diff --git a/collections/documentation/dashboard/dashboard.md b/collections/documentation/dashboard/dashboard.md
new file mode 100644
index 0000000..991b189
--- /dev/null
+++ b/collections/documentation/dashboard/dashboard.md
@@ -0,0 +1,43 @@
+
ThreeFold Dashboard
+
+Explore, control, and manage your ThreeFold Grid resources effortlessly through our integrated Dashboard. Deploy solutions seamlessly while gaining full control, all within a unified interface.
+
+The ThreeFold Dashboard is a revolutionary platform that simplifies the deployment process, allowing users to effortlessly interact with the TFGrid using intuitive web components known as weblets.
+
+## What is the ThreeFold Dashboard?
+
+The ThreeFold Dashboard is a dynamic environment designed for both seasoned developers and newcomers alike. It offers a seamless and accessible browser experience, making it easy to deploy solutions on the TFGrid through the use of weblets.
+
+In the context of the Dashboard, a weblet is a compiled JavaScript web component that can be effortlessly embedded within the HTML page of a web application. This modular approach allows for flexible and intuitive interactions, facilitating a user-friendly deployment process.
+
+The backend for the weblets is introduced with the [Javascript Client](../developers/javascript/grid3_javascript_readme.md) which communicates to TFChain over RMB.
+
+ Table of Contents
+
+- [Wallet Connector](./wallet_connector.md)
+- [TFGrid](./tfgrid/tfgrid.md)
+- [Deploy](./deploy/deploy.md)
+- [Farms](./farms/farms.md)
+- [TFChain](./tfchain/tfchain.md)
+
+## Advantages
+
+- It is a non-code easy way to deploy a whole solution on the TFGrid.
+- It is 100% decentralized, there is no server involved.
+- It is powerful tool designed to empower individuals and organizations with seamless control and management over their ThreeFold Grid resources.
+- It provides an intuitive web-based interface that allows users to effortlessly deploy, monitor, and scale their workloads on the decentralized and sustainable ThreeFold Grid infrastructure.
+
+## Dashboard
+
+You can access the ThreeFold Dashboard on different TF Chain networks.
+
+- [https://dashboard.dev.grid.tf](https://dashboard.dev.grid.tf) for Dev net
+- [https://dashboard.qa.grid.tf](https://dashboard.qa.grid.tf) for QA net
+- [https://dashboard.test.grid.tf](https://dashboard.test.grid.tf) for Test net
+- [https://dashboard.grid.tf](https://dashboard.grid.tf) for Main net
+
+## Limitations
+
+- Regarding browser support, we're only supporting Google Chrome browser (and thus Brave browser) at the moment with more browsers to be supported soon.
+- Deploys one thing at a time.
+- Might take sometime to deploy a solution like Peertube, so you should wait a little bit until it's fully running.
diff --git a/collections/documentation/dashboard/deploy/applications.md b/collections/documentation/dashboard/deploy/applications.md
new file mode 100644
index 0000000..ca16e19
--- /dev/null
+++ b/collections/documentation/dashboard/deploy/applications.md
@@ -0,0 +1,24 @@
+ Ready Community Applications
+
+Easily deploy your favourite applications on the ThreeFold grid with a click of a button.
+
+![](../img/applications_landing.png)
+
+***
+
+ Table of Contents
+
+- [Algorand](../solutions/algorand.md)
+- [CasperLabs](../solutions/casper.md)
+- [Discourse](../solutions/discourse.md)
+- [Funkwhale](../solutions/funkwhale.md)
+- [Mattermost](../solutions/mattermost.md)
+- [Nextcloud](../solutions/nextcloud.md)
+- [Node Pilot](../solutions/nodepilot.md)
+- [ownCloud](../solutions/owncloud.md)
+- [Peertube](../solutions/peertube.md)
+- [Presearch](../solutions/presearch.md)
+- [Subsquid](../solutions/subsquid.md)
+- [Taiga](../solutions/taiga.md)
+- [Umbrel](../solutions/umbrel.md)
+- [WordPress](../solutions/wordpress.md)
\ No newline at end of file
diff --git a/collections/documentation/dashboard/deploy/dedicated_machines.md b/collections/documentation/dashboard/deploy/dedicated_machines.md
new file mode 100644
index 0000000..24f4dbb
--- /dev/null
+++ b/collections/documentation/dashboard/deploy/dedicated_machines.md
@@ -0,0 +1,93 @@
+ Dedicated Machines
+
+ Table of Contents
+
+- [What is a Dedicated Machine?](#what-is-a-dedicated-machine)
+- [Description](#description)
+- [Billing \& Pricing](#billing--pricing)
+- [Discounts](#discounts)
+- [Usage](#usage)
+- [GPU Support](#gpu-support)
+- [Filter and Reserve a GPU Node](#filter-and-reserve-a-gpu-node)
+ - [Filter Nodes](#filter-nodes)
+ - [Reserve a Node](#reserve-a-node)
+- [GPU Support Links](#gpu-support-links)
+
+***
+
+## What is a Dedicated Machine?
+
+Dedicated machines are 3Nodes that can be reserved and rented entirely by one user. The user can thus reserve an entire node and use it exclusively to deploy solutions. This feature is ideal for users who want to host heavy deployments with the benefits of high reliability and cost effectiveness.
+
+## Description
+
+- Node reserved with deploying a `RentContract` on this node. node can has only one rentContract.
+- When a user create a RentContract against a node, the grid validate that there are no other active contracts on that node on the creation.
+- Once a RentContract is created, the grid can only accept contracts on this node from the tenant.
+- Only workloads from the tenant are accepted
+
+## Billing & Pricing
+
+- Once a node is rented, there is a fixed charge billed to the tenant regardless of deployed workloads.
+- Any subsequent NodeContract deployed on a node where a rentContract is active (and the same user is creating the nodeContracts) can be excluded from billing (apart from public ip and network usage).
+- Billing rates are calculated hourly on the TFGrid.
+ - While some of the documentation mentions a monthly price, the chain expresses pricing per hour. The monthly price shown within the manual is offered as a convenience to users, as it provides a simple way to estimate costs.
+
+## Discounts
+
+- Received Discounts for renting a node on TFGrid internet capacity
+ - 50% for dedicated node (TF Pricing policies)
+ - A second level discount up to 60% for balance level see [Discount Levels](../../../knowledge_base/cloud/pricing/staking_discount_levels.md)
+- Discounts are calculated every time the grid bills by checking the available TFT balance on the user wallet and seeing if it is sufficient to receive a discount. As a result, if the user balance drops below the treshold of a given discount, the deployment price increases.
+
+## Usage
+
+- See list of all dedicated node on `Dedicated Machines` tab on the portal.
+
+ ![ ](../img/dedicated_machines.png)
+
+ - Hover over the price to see the applied discounts
+
+ ![](../img/dashboard_dedicated_nodes_discounts.png)
+
+ - Expand row to see more info on the node:
+
+ ![ ](../img/dashboard_dedicated_nodes_details.png)
+ - Resources
+ - Location
+ - Possible Public Ips *this depends on the farm it belongs to*
+
+ - You can see the nodes in 2 states:
+ - Free
+ - Reserved *Owned by current twin*
+- Reserve a node:
+ - If node is not rented by another twin you can simply click reserve.
+
+
+- Unreserve a node:
+ - Simply as reserving but another check will be done to check you don't have any active workloads on the node before unreserving.
+
+## GPU Support
+
+To use a GPU on the TFGrid, users need to rent a dedicated node. Once they have rented a dedicated node equipped with a GPU, users can deploy workloads on their dedicated GPU node.
+
+## Filter and Reserve a GPU Node
+
+You can filter and reserve a GPU node using the [Dedicated Machines section](https://dashboard.grid.tf/#/deploy/dedicated-nodes/) of the **Dashboard**.
+
+### Filter Nodes
+
+- Filter nodes using the vendor name
+ - In **Filters**, select **GPU's vendor name**
+ - Write the name of the vendor desired (e.g. **nvidia**, **amd**)
+- Filter nodes using the device name
+ - In **Filters**, select **GPU's device name**
+ - Write the name of the device desired (e.g. **GT218**)
+
+### Reserve a Node
+
+When you have decided which node to reserve, click on **Reserve** under the column named **Actions**. Once you've rented a dedicated node that has a GPU, you can deploy GPU workloads.
+
+## GPU Support Links
+
+The ThreeFold Manual covers many ways to use a GPU node on the TFGrid. Read [this section](../../system_administrators/gpu/gpu_toc.md) to learn more.
\ No newline at end of file
diff --git a/collections/documentation/dashboard/deploy/deploy.md b/collections/documentation/dashboard/deploy/deploy.md
new file mode 100644
index 0000000..a96decc
--- /dev/null
+++ b/collections/documentation/dashboard/deploy/deploy.md
@@ -0,0 +1,27 @@
+# Deploy
+
+Here you will find everything related to deployments on the ThreeFold grid. This includes:
+
+- Checking the cost of a deployment using [Pricing Calculator](./pricing_calculator.md)
+- Finding a node to deploy on using the [Node Finder](./node_finder.md)
+- Deploying your desired workload from [Virtual Machines](../solutions/vm_intro.md), [Orchestrators](./orchestrators.md), or [Applictions](./applications.md)
+- Renting your own node on the ThreeFold grid from [Dedicated Machines](./dedicated_machines.md)
+- Consulting [Your Contracts](./your_contracts.md) on the TFGrid
+- Finding or publishing Flists from [Images](./images.md)
+- Updating or generating your SSH key from [SSH Keys](./ssh_keys.md)
+
+ ![](../img/sidebar_2.png)
+
+***
+
+## Table of Content
+
+- [Pricing Calculator](./pricing_calculator.md)
+- [Node Finder](./node_finder.md)
+- [Virtual Machines](../solutions/vm_intro.md)
+- [Orchestrators](./orchestrators.md)
+- [Dedicated Machines](./dedicated_machines.md)
+- [Applications](./applications.md)
+- [Your Contracts](./your_contracts.md)
+- [Images](./images.md)
+- [SSH Keys](./ssh_keys.md)
\ No newline at end of file
diff --git a/collections/documentation/dashboard/deploy/images.md b/collections/documentation/dashboard/deploy/images.md
new file mode 100644
index 0000000..2c27966
--- /dev/null
+++ b/collections/documentation/dashboard/deploy/images.md
@@ -0,0 +1,5 @@
+# Images
+
+Find or Publish your Flist from [Zero-OS Hub](https://hub.grid.tf/)
+
+![](../img/0_hub.png)
\ No newline at end of file
diff --git a/collections/documentation/dashboard/deploy/node_finder.md b/collections/documentation/dashboard/deploy/node_finder.md
new file mode 100644
index 0000000..dd93347
--- /dev/null
+++ b/collections/documentation/dashboard/deploy/node_finder.md
@@ -0,0 +1,40 @@
+Node Finder
+
+Table of Contents
+
+- [Nodes](#nodes)
+- [GPU Support](#gpu-support)
+
+***
+
+## Nodes
+
+The Node Finder page provides a more detailed view for the nodes available on the ThreeFold grid With detailed information and statistics about any of the available nodes.
+
+![](../img/nodes.png)
+
+You can get a node with the desired specifications using the filters available in the nodes page.
+
+![](../img/nodes_filters.png)
+
+You can see all of the node details by clicking on a node record.
+
+![](../img/nodes_details.png)
+
+## GPU Support
+
+![GPU support](../img/gpu_filter.png)
+
+- A new filter for GPU supported node is now available on the Nodes page.
+- GPU count
+- Filtering capabilities based on the model / device
+
+On the details pages is shown the card information and its status (`reserved` or `available`) also the ID that’s needed to be used during deployments is easily accessible and has a copy to clipboard button.
+
+![GPU details](../img/gpu_details.png)
+
+Here’s an example of how it looks in case of reserved
+
+![GPU details](../img/gpu_details_reserved.png)
+
+The TF Dashboard is where to reserve the nodes the farmer should be able to set the extra fees on the form and the user also should be able to reserve and get the details of the node (cost including the extrafees, GPU informations).
diff --git a/collections/documentation/dashboard/deploy/node_finder_gpu_support.md b/collections/documentation/dashboard/deploy/node_finder_gpu_support.md
new file mode 100644
index 0000000..7002630
--- /dev/null
+++ b/collections/documentation/dashboard/deploy/node_finder_gpu_support.md
@@ -0,0 +1,19 @@
+ GPU Support
+
+***
+
+![GPU support](../img/gpu_filter.png)
+
+- A new filter for GPU supported node is now available on the Nodes page.
+- GPU count
+- Filtering capabilities based on the model / device
+
+On the details pages is shown the card information and its status (`reserved` or `available`) also the ID that’s needed to be used during deployments is easily accessible and has a copy to clipboard button.
+
+![GPU details](../img/gpu_details.png)
+
+Here’s an example of how it looks in case of reserved
+
+![GPU details](../img/gpu_details_reserved.png)
+
+The TF Dashboard is where to reserve the nodes the farmer should be able to set the extra fees on the form and the user also should be able to reserve and get the details of the node (cost including the extrafees, GPU informations).
diff --git a/collections/documentation/dashboard/deploy/orchestrators.md b/collections/documentation/dashboard/deploy/orchestrators.md
new file mode 100644
index 0000000..47282a6
--- /dev/null
+++ b/collections/documentation/dashboard/deploy/orchestrators.md
@@ -0,0 +1,14 @@
+# Orchestrators
+
+Deploy your favorite orchestrating services and enjoy the seamless coordination and automation of various software applications and services.
+
+![](../img/orchestrator_landing.png)
+
+***
+
+## Table of Contnet
+
+- [Kubernetes](../solutions/k8s.md)
+- [Caprover](../solutions/caprover.md)
+ - [Caprover Admin](../solutions/caprover_admin.md)
+ - [Caprover Worker](../solutions/caprover_worker.md)
\ No newline at end of file
diff --git a/collections/documentation/dashboard/deploy/pricing_calculator.md b/collections/documentation/dashboard/deploy/pricing_calculator.md
new file mode 100644
index 0000000..a050215
--- /dev/null
+++ b/collections/documentation/dashboard/deploy/pricing_calculator.md
@@ -0,0 +1,6 @@
+# TF Resource Calculator
+
+A tool provided by ThreeFold that allows users to estimate and calculate potential cost of a deployment on the ThreeFold grid. The resource calculator takes into account various factors, including deployment resources, node certification, currnet balance, and in return it displays an accurate estimate for the deployment in terms of ThreeFold Tokens (TFT) and in USD per month.
+
+
+![](../img/pricing_calculator.png)
\ No newline at end of file
diff --git a/collections/documentation/dashboard/deploy/ssh_keys.md b/collections/documentation/dashboard/deploy/ssh_keys.md
new file mode 100644
index 0000000..4e2cecb
--- /dev/null
+++ b/collections/documentation/dashboard/deploy/ssh_keys.md
@@ -0,0 +1,5 @@
+# SSH Keys
+
+Add, update or generate your SSH key with a click of a button.
+
+![](../img/SSH_Key.png)
\ No newline at end of file
diff --git a/collections/documentation/dashboard/deploy/vm.md b/collections/documentation/dashboard/deploy/vm.md
new file mode 100644
index 0000000..5a22576
--- /dev/null
+++ b/collections/documentation/dashboard/deploy/vm.md
@@ -0,0 +1,13 @@
+ Virtual Machines
+
+On the TFGrid, you can deploy both micro and full virtual machines.
+
+![](../img/vm_landing.png)
+
+***
+
+ Table of Contents
+
+- [Micro and Full VM Differences ](../solutions/vm_differences.md)
+- [Full Virtual Machine](../solutions/fullVm.md)
+- [Micro Virtual Machine](../solutions/vm.md)
diff --git a/collections/documentation/dashboard/deploy/your_contracts.md b/collections/documentation/dashboard/deploy/your_contracts.md
new file mode 100644
index 0000000..c7cb4a6
--- /dev/null
+++ b/collections/documentation/dashboard/deploy/your_contracts.md
@@ -0,0 +1,49 @@
+# Contracts
+
+From the Contracts section you can check your contracts by navigating to the `Deploy` then `Your Contracts` tab from the sidebar.
+
+From there you will see the `Contracts List`, the list is split into three different sections. these sections are:
+
+### Node contracts
+
+![image](../img/node_contracts.png)
+
+### Name contracts
+
+![image](../img/name_contracts.png)
+
+### Rent contracts
+
+![image](../img/rent_contracts.png)
+
+
+
+This list includes the following information about each contract.
+
+- Contract ID.
+- Contract Type.
+- Contract State (Created, Deleted, GracePeriod).
+- Solution Typw
+- Billing Rate (in TFT/Hour).
+- Solution Name.
+- Created At.
+- Expiration (Only appears if the contract is in GracePeriod).
+- Node ID
+- Node Status (Up, Down, Standby).
+- Show Details (This button will display the detailed information of the desired contract).
+
+ ![image](../img/contract_details.png)
+
+
+## Cancel Contract
+
+You can also cancel the target contract/s by select the contract you want to cancel
+
+- Click on the checkbox on the left side of the contract row
+- Click on the delete button in the bottom right of the table
+- Review the contract/s ID then click on *Delete* button
+
+Note:
+
+>- You can Cancel all you contracts by clicking on the checkbox on the left side of the table header then click on *Delete* button.
+>- It is advisable to remove the contract from its solution page, especially when multiple contracts may be linked to the same instance.
diff --git a/collections/documentation/dashboard/farms/farms.md b/collections/documentation/dashboard/farms/farms.md
new file mode 100644
index 0000000..7faedd3
--- /dev/null
+++ b/collections/documentation/dashboard/farms/farms.md
@@ -0,0 +1,19 @@
+# Farms
+
+Here you will find everything farming related. this includes:
+
+- Monitoring, creating, and updating your farms from the [Your Farms](./your_farms.md) section where you can also check your nodes and update multiple things like the public configuration and extra fees of the node.
+- Exploring and finding farms that are available on the ThreeFold grid from the [Farm Finder](./farms_finder.md) section.
+- Generating your own boot device for your system from the [Node Installer](./node_installer.md) section.
+- Estimating and calculating potential earnings from farming on the ThreeFold Grid from the [Simulator](./simulator.md) section.
+
+ ![](../img/sidebar_3.png)
+
+***
+
+## Table of Content
+
+- [Your Farms](./your_farms.md)
+- [Farm Finder](./farms_finder.md)
+- [Node Installer](./node_installer.md)
+- [Simulator](./simulator.md)
\ No newline at end of file
diff --git a/collections/documentation/dashboard/farms/farms_finder.md b/collections/documentation/dashboard/farms/farms_finder.md
new file mode 100644
index 0000000..252aa73
--- /dev/null
+++ b/collections/documentation/dashboard/farms/farms_finder.md
@@ -0,0 +1,13 @@
+# Farms
+
+The farms page provides a more detailed view for the farms available on the ThreeFold grid With detailed information about any of the available farms.
+
+![](../img/farms.png)
+
+You can search for a specific farm using the farms filters.
+
+![](../img/farms_filters.png)
+
+You can see all of the farm details by clicking on a farm record.
+
+![](../img/farms_details.png)
diff --git a/collections/documentation/dashboard/farms/node_installer.md b/collections/documentation/dashboard/farms/node_installer.md
new file mode 100644
index 0000000..903e10e
--- /dev/null
+++ b/collections/documentation/dashboard/farms/node_installer.md
@@ -0,0 +1,5 @@
+# Node Installer
+
+Generate your own boot device for your system and download Zero-OS Images from [Zero-OS Bootstrap](https://bootstrap.grid.tf/)
+
+![](../img/0_Bootstrap.png)
\ No newline at end of file
diff --git a/collections/documentation/dashboard/farms/simulator.md b/collections/documentation/dashboard/farms/simulator.md
new file mode 100644
index 0000000..3d5b594
--- /dev/null
+++ b/collections/documentation/dashboard/farms/simulator.md
@@ -0,0 +1,5 @@
+# Simulator
+
+A tool provided by ThreeFold that allows users to estimate and calculate potential earnings from farming on the ThreeFold Grid. Farming refers to the process of providing computing resources, such as storage and processing power, to the ThreeFold Grid and earning tokens in return. The tf-farming-calculator takes into account various factors, including the amount of resources contributed, the duration of farming, and the current market conditions, to provide users with an estimate of their potential earnings in terms of ThreeFold Tokens (TFT).
+
+![](../img/simulator.png)
\ No newline at end of file
diff --git a/collections/documentation/dashboard/farms/your_farms.md b/collections/documentation/dashboard/farms/your_farms.md
new file mode 100644
index 0000000..8afc756
--- /dev/null
+++ b/collections/documentation/dashboard/farms/your_farms.md
@@ -0,0 +1,134 @@
+# Farms
+
+This comprehensive guide aims to provide users with detailed instructions and insights into efficiently managing their _Farms_. Farms encompass servers and storage devices contributing computational and storage capabilities to the grid, empowering users to oversee, maintain, and optimize their resources effectively.
+
+- [Getting started](#getting-started)
+- [Create a new Farm](#create-a-new-farm)
+- [Manage Your Farms](#manage-your-farms)
+ - [Add a public IP to your Farm](#add-a-public-ip-to-your-farm)
+ - [Add a Stellar address for payout](#add-a-stellar-address-for-payout)
+ - [Generate your node bootstrap image](#generate-your-node-bootstrap-image)
+ - [Additional information](#additional-information)
+- [Manage Your Nodes](#manage-your-nodes)
+ - [Node information](#node-information)
+ - [Extra Fees](#extra-fees)
+ - [Public Configuration](#public-configuration)
+ - [The Difference Between IPs Assigned to Nodes Versus a Farm](#the-difference-between-ips-assigned-to-nodes-versus-a-farm)
+
+## Getting started
+
+After logging in to the TF Dashboard, on the sidebar click on **Dashboard** then _Your Farms_ .
+
+## Create a new Farm
+
+If you want to start farming, you need a farmID, the ID of the farm that is owning the hardware node(s) you connect to the TFGrid.
+
+**Currently on**:
+
+- [Devnet](https://dashboard.dev.grid.tf/)
+- [Qanet](https://dashboard.qa.grid.tf/)
+- [Testnet](https://dashboard.test.grid.tf/)
+- [Mainnet](https://dashboard.grid.tf/)
+
+Click `Create Farm` and choose a name.
+
+![ ](../img/dashboard_farms.png)
+
+![ ](../img/dashboard_farms_create.png)
+
+Click on `Create`.
+
+The farm is by default set up as 'DIY'. A farm can become certified through certification program.
+Also a pricing policy is defined. Pricing policy is currently the same for all farms, the field is created for future use.
+
+## Manage Your Farms
+
+You can browse your Farms in _Farms_ table; Farms table contains all your own farms and its your entry point to manage your farm as in the following sections.
+
+![](../img/dashboard_farms_farms_table.png)
+
+### Add a public IP to your Farm
+
+If you have public IPv4 addresses available that can be used for usage on the TFGrid, you can add them in your farm.
+Click `ADD IP`, specify the addresses, the gateway and click `CREATE`.
+You can add them one by one or using range of IPs.
+
+**Some notes about adding a new IPs**:
+
+- Be careful not to create a new IP range that contains an IP address that already exists; doing so will result in an error.
+- Verify that both the gateway address and the IP address are correct.
+- Be careful not to include the same gateway address in a new IP range.
+
+![ ](../img/dashboard_farms_farm_details.png)
+
+![ ](../img/dashboard_farms_add_ip_single.png)
+
+![ ](../img/dashboard_farms_add_ip_range.png)
+
+Deleting IPv4 addresses is also possible here. The `Deployed Contract ID` gives an indication of whether an IP is currently used. If it is 0, it is safe to remove it.
+
+![ ](../img/dashboard_farms_ip_details.png)
+
+### Add a Stellar address for payout
+
+In a first phase, farming of tokens still results in payout on the Stellar network. So to get the farming reward, a Stellar address needs to be provided.
+
+![ ](../img/dashboard_farms_farm_details.png)
+
+![ ](../img/dashboard_farms_stellar_address.png)
+
+You can read about different ways to store TFT [here](../../threefold_token/storing_tft/storing_tft.md). Make sure to use a Stellar wallet for your farming rewards.
+
+### Generate your node bootstrap image
+
+Once you know your farmID, you can set up your node on TFGrid3. Click on `Bootstrap Node Image`.
+
+![dashboard_bootstrap_farm](../img/dashboard_bootstrap_farm.png)
+
+Read more Zero-OS bootstrap image [here](../../farmers/3node_building/2_bootstrap_image.md).
+
+### Additional information
+
+After booting a node, the info will become available in `Your Nodes` table, including the status info along with the minting and fixup receipts.
+
+![ ](../img/dashboard_farms_node_details.png)
+
+Clicking on the node statistics will open up a calendar where you can view the periods the node was minting or undergoing a fixup. Clicking on the periods will show a popup with the start and end datetimes, receipt hash and the amount of TFTs minted (if it is a minting receipt).
+
+![ ](../img/dashboard_portal_ui_nodes_minting.png)
+
+You can also download a single node's receipts using the `Download Receipts` button within the node statistics. Moreover, you can download all of the nodes' receipts using the `Download Receipts` button on the top left corner of the farm nodes table.
+
+## Manage Your Nodes
+
+as in farms table _Nodes_ table contains all your own nodes and its your entry point to manage your farm as in the following sections.
+
+### Node information
+
+Expand your node information by clicking on the expand button in the target node row.
+
+### Extra Fees
+
+You can set a price for the special hardware you’re providing e.g. GPUs while renting.
+
+![](../img/dashboard_farms_extra_fee.png)
+
+- Under the **Your Nodes** table, locate the target node and click **Set Additional Fees** under **Actions**
+- Set a monthly fee (in USD) and click **Save**
+
+### Public Configuration
+
+To configure public IP addresses to a specific Node
+
+![](../img/dashboard_farms_public_config.png)
+
+- Under the **Your Nodes** table, locate the target node and click **Add a public config** under **Actions**
+- Fill in the necessary information and click save. Only the IPv4 address and gateway are necessary.
+
+> The IPv6 address and the Domain are optional but if you provide The IPv6 you have to provide its Domain.
+
+#### The Difference Between IPs Assigned to Nodes Versus a Farm
+
+---
+
+IPs assigned to a farm are available to be rented by workloads. They can be assigned to virtual machines for example. IPs assigned to nodes enable each node to become a gateway.
diff --git a/collections/documentation/dashboard/home.md b/collections/documentation/dashboard/home.md
new file mode 100644
index 0000000..d14fd9a
--- /dev/null
+++ b/collections/documentation/dashboard/home.md
@@ -0,0 +1,43 @@
+ ThreeFold Dashboard
+
+Explore, control, and manage your ThreeFold Grid resources effortlessly through our integrated Dashboard. Deploy solutions seamlessly while gaining full control, all within a unified interface.
+
+The ThreeFold Dashboard is a revolutionary platform that simplifies the deployment process, allowing users to effortlessly interact with the TFGrid using intuitive web components known as weblets.
+
+## What is the ThreeFold Dashboard?
+
+The ThreeFold Dashboard is a dynamic environment designed for both seasoned developers and newcomers alike. It offers a seamless and accessible browser experience, making it easy to deploy solutions on the TFGrid through the use of weblets.
+
+In the context of the Dashboard, a weblet is a compiled JavaScript web component that can be effortlessly embedded within the HTML page of a web application. This modular approach allows for flexible and intuitive interactions, facilitating a user-friendly deployment process.
+
+The backend for the weblets is introduced with the [Javascript Client](../javascript/grid3_javascript_readme.md) which communicates to TFChain over RMB.
+
+ Table of Contents
+
+- [Wallet Connector](./wallet_connector.md)
+- [TFGrid](./tfgrid/tfgrid.md)
+- [Deploy](./deploy/deploy.md)
+- [Farms](./farms/farms.md)
+- [TFChain](./tfchain/tfchain.md)
+
+## Advantages
+
+- It is a non-code easy way to deploy a whole solution on the TFGrid.
+- It is 100% decentralized, there is no server involved.
+- It is powerful tool designed to empower individuals and organizations with seamless control and management over their ThreeFold Grid resources.
+- It provides an intuitive web-based interface that allows users to effortlessly deploy, monitor, and scale their workloads on the decentralized and sustainable ThreeFold Grid infrastructure.
+
+## Dashboard
+
+You can access the ThreeFold Dashboard on different TF Chain networks.
+
+- [https://dashboard.dev.grid.tf](https://dashboard.dev.grid.tf) for Dev net.
+- [https://dashboard.qa.grid.tf](https://dashboard.qa.grid.tf) for QA net.
+- [https://dashboard.test.grid.tf](https://dashboard.test.grid.tf) for Test net.
+- [https://dashboard.grid.tf](https://dashboard.grid.tf) for Main net.
+
+## Limitations
+
+- Regarding browser support, we're only supporting Google Chrome browser (and thus Brave browser) at the moment with more browsers to be supported soon.
+- Deploys one thing at a time.
+- Might take sometime to deploy a solution like Peertube, so you should wait a little bit until it's fully running.
diff --git a/collections/documentation/dashboard/img/0_Bootstrap.png b/collections/documentation/dashboard/img/0_Bootstrap.png
new file mode 100644
index 0000000..dedec1b
Binary files /dev/null and b/collections/documentation/dashboard/img/0_Bootstrap.png differ
diff --git a/collections/documentation/dashboard/img/0_hub.png b/collections/documentation/dashboard/img/0_hub.png
new file mode 100644
index 0000000..da7faf5
Binary files /dev/null and b/collections/documentation/dashboard/img/0_hub.png differ
diff --git a/collections/documentation/dashboard/img/Minting.png b/collections/documentation/dashboard/img/Minting.png
new file mode 100644
index 0000000..b7de134
Binary files /dev/null and b/collections/documentation/dashboard/img/Minting.png differ
diff --git a/collections/documentation/dashboard/img/Monitoring.png b/collections/documentation/dashboard/img/Monitoring.png
new file mode 100644
index 0000000..1cda446
Binary files /dev/null and b/collections/documentation/dashboard/img/Monitoring.png differ
diff --git a/collections/documentation/dashboard/img/SSH_Key.png b/collections/documentation/dashboard/img/SSH_Key.png
new file mode 100644
index 0000000..7d22832
Binary files /dev/null and b/collections/documentation/dashboard/img/SSH_Key.png differ
diff --git a/collections/documentation/dashboard/img/add_domain_10.png b/collections/documentation/dashboard/img/add_domain_10.png
new file mode 100644
index 0000000..aabe5de
Binary files /dev/null and b/collections/documentation/dashboard/img/add_domain_10.png differ
diff --git a/collections/documentation/dashboard/img/add_domain_11.png b/collections/documentation/dashboard/img/add_domain_11.png
new file mode 100644
index 0000000..8540ce8
Binary files /dev/null and b/collections/documentation/dashboard/img/add_domain_11.png differ
diff --git a/collections/documentation/dashboard/img/add_domain_6.png b/collections/documentation/dashboard/img/add_domain_6.png
new file mode 100644
index 0000000..05611ef
Binary files /dev/null and b/collections/documentation/dashboard/img/add_domain_6.png differ
diff --git a/collections/documentation/dashboard/img/add_domain_8.png b/collections/documentation/dashboard/img/add_domain_8.png
new file mode 100644
index 0000000..3a35803
Binary files /dev/null and b/collections/documentation/dashboard/img/add_domain_8.png differ
diff --git a/collections/documentation/dashboard/img/add_domain_9.png b/collections/documentation/dashboard/img/add_domain_9.png
new file mode 100644
index 0000000..346009b
Binary files /dev/null and b/collections/documentation/dashboard/img/add_domain_9.png differ
diff --git a/collections/documentation/dashboard/img/add_new_domain_success.png b/collections/documentation/dashboard/img/add_new_domain_success.png
new file mode 100644
index 0000000..3b23776
Binary files /dev/null and b/collections/documentation/dashboard/img/add_new_domain_success.png differ
diff --git a/collections/documentation/dashboard/img/applications_landing.png b/collections/documentation/dashboard/img/applications_landing.png
new file mode 100644
index 0000000..e4117a5
Binary files /dev/null and b/collections/documentation/dashboard/img/applications_landing.png differ
diff --git a/collections/documentation/dashboard/img/bridge.png b/collections/documentation/dashboard/img/bridge.png
new file mode 100644
index 0000000..269dead
Binary files /dev/null and b/collections/documentation/dashboard/img/bridge.png differ
diff --git a/collections/documentation/dashboard/img/bridge_deposit.png b/collections/documentation/dashboard/img/bridge_deposit.png
new file mode 100644
index 0000000..97619cc
Binary files /dev/null and b/collections/documentation/dashboard/img/bridge_deposit.png differ
diff --git a/collections/documentation/dashboard/img/bridge_withdraw.png b/collections/documentation/dashboard/img/bridge_withdraw.png
new file mode 100644
index 0000000..461a80d
Binary files /dev/null and b/collections/documentation/dashboard/img/bridge_withdraw.png differ
diff --git a/collections/documentation/dashboard/img/contract_details.png b/collections/documentation/dashboard/img/contract_details.png
new file mode 100644
index 0000000..286c87a
Binary files /dev/null and b/collections/documentation/dashboard/img/contract_details.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_T&C.png b/collections/documentation/dashboard/img/dashboard_T&C.png
new file mode 100644
index 0000000..6173f1a
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_T&C.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_bootstrap_farm.png b/collections/documentation/dashboard/img/dashboard_bootstrap_farm.png
new file mode 100644
index 0000000..82f8759
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_bootstrap_farm.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_bridge.png b/collections/documentation/dashboard/img/dashboard_bridge.png
new file mode 100644
index 0000000..111ef91
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_bridge.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_connect.png b/collections/documentation/dashboard/img/dashboard_connect.png
new file mode 100644
index 0000000..289b2e8
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_connect.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_dao.png b/collections/documentation/dashboard/img/dashboard_dao.png
new file mode 100644
index 0000000..37fb232
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_dao.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_dedicated_nodes_details.png b/collections/documentation/dashboard/img/dashboard_dedicated_nodes_details.png
new file mode 100644
index 0000000..649f7c9
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_dedicated_nodes_details.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_dedicated_nodes_discounts.png b/collections/documentation/dashboard/img/dashboard_dedicated_nodes_discounts.png
new file mode 100644
index 0000000..fe66434
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_dedicated_nodes_discounts.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_explorer_farms.png b/collections/documentation/dashboard/img/dashboard_explorer_farms.png
new file mode 100644
index 0000000..58d3305
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_explorer_farms.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_explorer_nodes.png b/collections/documentation/dashboard/img/dashboard_explorer_nodes.png
new file mode 100644
index 0000000..963730e
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_explorer_nodes.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_explorer_statistics.png b/collections/documentation/dashboard/img/dashboard_explorer_statistics.png
new file mode 100644
index 0000000..0d81f19
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_explorer_statistics.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_farms.png b/collections/documentation/dashboard/img/dashboard_farms.png
new file mode 100644
index 0000000..bebffe2
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_farms.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_farms_add_ip_range.png b/collections/documentation/dashboard/img/dashboard_farms_add_ip_range.png
new file mode 100644
index 0000000..30d6dcc
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_farms_add_ip_range.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_farms_add_ip_single.png b/collections/documentation/dashboard/img/dashboard_farms_add_ip_single.png
new file mode 100644
index 0000000..26f80e4
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_farms_add_ip_single.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_farms_create.png b/collections/documentation/dashboard/img/dashboard_farms_create.png
new file mode 100644
index 0000000..684dfbb
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_farms_create.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_farms_extra_fee.png b/collections/documentation/dashboard/img/dashboard_farms_extra_fee.png
new file mode 100644
index 0000000..2800f03
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_farms_extra_fee.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_farms_farm_details.png b/collections/documentation/dashboard/img/dashboard_farms_farm_details.png
new file mode 100644
index 0000000..5de107e
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_farms_farm_details.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_farms_farms_table.png b/collections/documentation/dashboard/img/dashboard_farms_farms_table.png
new file mode 100644
index 0000000..dd92e0b
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_farms_farms_table.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_farms_ip_details.png b/collections/documentation/dashboard/img/dashboard_farms_ip_details.png
new file mode 100644
index 0000000..025d9b5
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_farms_ip_details.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_farms_node_details.png b/collections/documentation/dashboard/img/dashboard_farms_node_details.png
new file mode 100644
index 0000000..c95e3fc
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_farms_node_details.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_farms_public_config.png b/collections/documentation/dashboard/img/dashboard_farms_public_config.png
new file mode 100644
index 0000000..cf1e8f5
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_farms_public_config.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_farms_stellar_address.png b/collections/documentation/dashboard/img/dashboard_farms_stellar_address.png
new file mode 100644
index 0000000..0ba6efd
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_farms_stellar_address.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_generate_account.png b/collections/documentation/dashboard/img/dashboard_generate_account.png
new file mode 100644
index 0000000..c0433dc
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_generate_account.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_login.png b/collections/documentation/dashboard/img/dashboard_login.png
new file mode 100644
index 0000000..c119ac3
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_login.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_portal_account.png b/collections/documentation/dashboard/img/dashboard_portal_account.png
new file mode 100644
index 0000000..be66b4e
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_portal_account.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_portal_create_account_1.png b/collections/documentation/dashboard/img/dashboard_portal_create_account_1.png
new file mode 100644
index 0000000..4024e7d
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_portal_create_account_1.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_portal_create_account_2.png b/collections/documentation/dashboard/img/dashboard_portal_create_account_2.png
new file mode 100644
index 0000000..5f1b4e5
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_portal_create_account_2.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_portal_create_twin.png b/collections/documentation/dashboard/img/dashboard_portal_create_twin.png
new file mode 100644
index 0000000..776a27e
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_portal_create_twin.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_portal_deposit_tft.png b/collections/documentation/dashboard/img/dashboard_portal_deposit_tft.png
new file mode 100644
index 0000000..97782e3
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_portal_deposit_tft.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_portal_farms.png b/collections/documentation/dashboard/img/dashboard_portal_farms.png
new file mode 100644
index 0000000..3272e64
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_portal_farms.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_portal_fill_ipv6.png b/collections/documentation/dashboard/img/dashboard_portal_fill_ipv6.png
new file mode 100644
index 0000000..c2ddd0c
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_portal_fill_ipv6.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_portal_terms_conditions.png b/collections/documentation/dashboard/img/dashboard_portal_terms_conditions.png
new file mode 100644
index 0000000..82f6c3e
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_portal_terms_conditions.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_portal_transaction_sign.png b/collections/documentation/dashboard/img/dashboard_portal_transaction_sign.png
new file mode 100644
index 0000000..62b4219
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_portal_transaction_sign.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_portal_transfer_detail.png b/collections/documentation/dashboard/img/dashboard_portal_transfer_detail.png
new file mode 100644
index 0000000..919303c
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_portal_transfer_detail.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_portal_twin_created.png b/collections/documentation/dashboard/img/dashboard_portal_twin_created.png
new file mode 100644
index 0000000..0d9b470
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_portal_twin_created.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_portal_ui_nodes_minting.png b/collections/documentation/dashboard/img/dashboard_portal_ui_nodes_minting.png
new file mode 100644
index 0000000..fc925e2
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_portal_ui_nodes_minting.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_portal_withdraw_tft.png b/collections/documentation/dashboard/img/dashboard_portal_withdraw_tft.png
new file mode 100644
index 0000000..5493694
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_portal_withdraw_tft.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_sidebar_portal.png b/collections/documentation/dashboard/img/dashboard_sidebar_portal.png
new file mode 100644
index 0000000..9639032
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_sidebar_portal.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_statistics.png b/collections/documentation/dashboard/img/dashboard_statistics.png
new file mode 100644
index 0000000..8badd4c
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_statistics.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_swap.png b/collections/documentation/dashboard/img/dashboard_swap.png
new file mode 100644
index 0000000..dbda2ba
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_swap.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_transfer_address.png b/collections/documentation/dashboard/img/dashboard_transfer_address.png
new file mode 100644
index 0000000..2a389e9
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_transfer_address.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_transfer_twin.png b/collections/documentation/dashboard/img/dashboard_transfer_twin.png
new file mode 100644
index 0000000..7bff8ca
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_transfer_twin.png differ
diff --git a/collections/documentation/dashboard/img/dashboard_uptime.png b/collections/documentation/dashboard/img/dashboard_uptime.png
new file mode 100644
index 0000000..21920e7
Binary files /dev/null and b/collections/documentation/dashboard/img/dashboard_uptime.png differ
diff --git a/collections/documentation/dashboard/img/dedicated_machines.png b/collections/documentation/dashboard/img/dedicated_machines.png
new file mode 100644
index 0000000..e832be5
Binary files /dev/null and b/collections/documentation/dashboard/img/dedicated_machines.png differ
diff --git a/collections/documentation/dashboard/img/explorer_basics_.png b/collections/documentation/dashboard/img/explorer_basics_.png
new file mode 100644
index 0000000..d8635ea
Binary files /dev/null and b/collections/documentation/dashboard/img/explorer_basics_.png differ
diff --git a/collections/documentation/dashboard/img/explorer_basics_2.png b/collections/documentation/dashboard/img/explorer_basics_2.png
new file mode 100644
index 0000000..8af783c
Binary files /dev/null and b/collections/documentation/dashboard/img/explorer_basics_2.png differ
diff --git a/collections/documentation/dashboard/img/explorer_darkmode.png b/collections/documentation/dashboard/img/explorer_darkmode.png
new file mode 100644
index 0000000..54cce34
Binary files /dev/null and b/collections/documentation/dashboard/img/explorer_darkmode.png differ
diff --git a/collections/documentation/dashboard/img/explorer_farm_details.png b/collections/documentation/dashboard/img/explorer_farm_details.png
new file mode 100644
index 0000000..17f5f88
Binary files /dev/null and b/collections/documentation/dashboard/img/explorer_farm_details.png differ
diff --git a/collections/documentation/dashboard/img/explorer_farms.png b/collections/documentation/dashboard/img/explorer_farms.png
new file mode 100644
index 0000000..73ddb75
Binary files /dev/null and b/collections/documentation/dashboard/img/explorer_farms.png differ
diff --git a/collections/documentation/dashboard/img/explorer_find_country_pubip.png b/collections/documentation/dashboard/img/explorer_find_country_pubip.png
new file mode 100644
index 0000000..54b7ab6
Binary files /dev/null and b/collections/documentation/dashboard/img/explorer_find_country_pubip.png differ
diff --git a/collections/documentation/dashboard/img/explorer_gpu.png b/collections/documentation/dashboard/img/explorer_gpu.png
new file mode 100644
index 0000000..6df7394
Binary files /dev/null and b/collections/documentation/dashboard/img/explorer_gpu.png differ
diff --git a/collections/documentation/dashboard/img/explorer_grafana.png b/collections/documentation/dashboard/img/explorer_grafana.png
new file mode 100644
index 0000000..9a702fb
Binary files /dev/null and b/collections/documentation/dashboard/img/explorer_grafana.png differ
diff --git a/collections/documentation/dashboard/img/explorer_node_details.png b/collections/documentation/dashboard/img/explorer_node_details.png
new file mode 100644
index 0000000..e6fd8c1
Binary files /dev/null and b/collections/documentation/dashboard/img/explorer_node_details.png differ
diff --git a/collections/documentation/dashboard/img/explorer_node_example_1.png b/collections/documentation/dashboard/img/explorer_node_example_1.png
new file mode 100644
index 0000000..712a8da
Binary files /dev/null and b/collections/documentation/dashboard/img/explorer_node_example_1.png differ
diff --git a/collections/documentation/dashboard/img/explorer_node_example_2.png b/collections/documentation/dashboard/img/explorer_node_example_2.png
new file mode 100644
index 0000000..7a0e46b
Binary files /dev/null and b/collections/documentation/dashboard/img/explorer_node_example_2.png differ
diff --git a/collections/documentation/dashboard/img/explorer_node_resources_1.png b/collections/documentation/dashboard/img/explorer_node_resources_1.png
new file mode 100644
index 0000000..51e44a0
Binary files /dev/null and b/collections/documentation/dashboard/img/explorer_node_resources_1.png differ
diff --git a/collections/documentation/dashboard/img/explorer_node_resources_2.png b/collections/documentation/dashboard/img/explorer_node_resources_2.png
new file mode 100644
index 0000000..0c406f3
Binary files /dev/null and b/collections/documentation/dashboard/img/explorer_node_resources_2.png differ
diff --git a/collections/documentation/dashboard/img/explorer_node_resources_3.png b/collections/documentation/dashboard/img/explorer_node_resources_3.png
new file mode 100644
index 0000000..47af607
Binary files /dev/null and b/collections/documentation/dashboard/img/explorer_node_resources_3.png differ
diff --git a/collections/documentation/dashboard/img/explorer_nodes.png b/collections/documentation/dashboard/img/explorer_nodes.png
new file mode 100644
index 0000000..d99a011
Binary files /dev/null and b/collections/documentation/dashboard/img/explorer_nodes.png differ
diff --git a/collections/documentation/dashboard/img/explorer_nodes_distribution.png b/collections/documentation/dashboard/img/explorer_nodes_distribution.png
new file mode 100644
index 0000000..90cd664
Binary files /dev/null and b/collections/documentation/dashboard/img/explorer_nodes_distribution.png differ
diff --git a/collections/documentation/dashboard/img/explorer_statistics.png b/collections/documentation/dashboard/img/explorer_statistics.png
new file mode 100644
index 0000000..9821a1d
Binary files /dev/null and b/collections/documentation/dashboard/img/explorer_statistics.png differ
diff --git a/collections/documentation/dashboard/img/explorer_status.png b/collections/documentation/dashboard/img/explorer_status.png
new file mode 100644
index 0000000..821d4ca
Binary files /dev/null and b/collections/documentation/dashboard/img/explorer_status.png differ
diff --git a/collections/documentation/dashboard/img/explorer_toggle_gateways.png b/collections/documentation/dashboard/img/explorer_toggle_gateways.png
new file mode 100644
index 0000000..6ce6c9f
Binary files /dev/null and b/collections/documentation/dashboard/img/explorer_toggle_gateways.png differ
diff --git a/collections/documentation/dashboard/img/explorer_toggle_gpu.png b/collections/documentation/dashboard/img/explorer_toggle_gpu.png
new file mode 100644
index 0000000..7ee4d6f
Binary files /dev/null and b/collections/documentation/dashboard/img/explorer_toggle_gpu.png differ
diff --git a/collections/documentation/dashboard/img/farms.png b/collections/documentation/dashboard/img/farms.png
new file mode 100644
index 0000000..5280633
Binary files /dev/null and b/collections/documentation/dashboard/img/farms.png differ
diff --git a/collections/documentation/dashboard/img/farms_details.png b/collections/documentation/dashboard/img/farms_details.png
new file mode 100644
index 0000000..744fde4
Binary files /dev/null and b/collections/documentation/dashboard/img/farms_details.png differ
diff --git a/collections/documentation/dashboard/img/farms_filters.png b/collections/documentation/dashboard/img/farms_filters.png
new file mode 100644
index 0000000..8eca69f
Binary files /dev/null and b/collections/documentation/dashboard/img/farms_filters.png differ
diff --git a/collections/documentation/dashboard/img/gpu_details.png b/collections/documentation/dashboard/img/gpu_details.png
new file mode 100644
index 0000000..83314cb
Binary files /dev/null and b/collections/documentation/dashboard/img/gpu_details.png differ
diff --git a/collections/documentation/dashboard/img/gpu_details_reserved.png b/collections/documentation/dashboard/img/gpu_details_reserved.png
new file mode 100644
index 0000000..84ab221
Binary files /dev/null and b/collections/documentation/dashboard/img/gpu_details_reserved.png differ
diff --git a/collections/documentation/dashboard/img/gpu_filter.png b/collections/documentation/dashboard/img/gpu_filter.png
new file mode 100644
index 0000000..3a3657d
Binary files /dev/null and b/collections/documentation/dashboard/img/gpu_filter.png differ
diff --git a/collections/documentation/dashboard/img/grid_health.png b/collections/documentation/dashboard/img/grid_health.png
new file mode 100644
index 0000000..2c5d9ee
Binary files /dev/null and b/collections/documentation/dashboard/img/grid_health.png differ
diff --git a/collections/documentation/dashboard/img/list_domains.png b/collections/documentation/dashboard/img/list_domains.png
new file mode 100644
index 0000000..9144a76
Binary files /dev/null and b/collections/documentation/dashboard/img/list_domains.png differ
diff --git a/collections/documentation/dashboard/img/manage_domains_button.png b/collections/documentation/dashboard/img/manage_domains_button.png
new file mode 100644
index 0000000..891110b
Binary files /dev/null and b/collections/documentation/dashboard/img/manage_domains_button.png differ
diff --git a/collections/documentation/dashboard/img/name_contracts.png b/collections/documentation/dashboard/img/name_contracts.png
new file mode 100644
index 0000000..35f1217
Binary files /dev/null and b/collections/documentation/dashboard/img/name_contracts.png differ
diff --git a/collections/documentation/dashboard/img/node_contracts.png b/collections/documentation/dashboard/img/node_contracts.png
new file mode 100644
index 0000000..b785b21
Binary files /dev/null and b/collections/documentation/dashboard/img/node_contracts.png differ
diff --git a/collections/documentation/dashboard/img/node_detail_.png b/collections/documentation/dashboard/img/node_detail_.png
new file mode 100644
index 0000000..1e06f9a
Binary files /dev/null and b/collections/documentation/dashboard/img/node_detail_.png differ
diff --git a/collections/documentation/dashboard/img/node_detail_1.png b/collections/documentation/dashboard/img/node_detail_1.png
new file mode 100644
index 0000000..6747192
Binary files /dev/null and b/collections/documentation/dashboard/img/node_detail_1.png differ
diff --git a/collections/documentation/dashboard/img/nodes.png b/collections/documentation/dashboard/img/nodes.png
new file mode 100644
index 0000000..be29183
Binary files /dev/null and b/collections/documentation/dashboard/img/nodes.png differ
diff --git a/collections/documentation/dashboard/img/nodes_details.png b/collections/documentation/dashboard/img/nodes_details.png
new file mode 100644
index 0000000..f5c4e8b
Binary files /dev/null and b/collections/documentation/dashboard/img/nodes_details.png differ
diff --git a/collections/documentation/dashboard/img/nodes_filters.png b/collections/documentation/dashboard/img/nodes_filters.png
new file mode 100644
index 0000000..68257c2
Binary files /dev/null and b/collections/documentation/dashboard/img/nodes_filters.png differ
diff --git a/collections/documentation/dashboard/img/orchestrator_landing.png b/collections/documentation/dashboard/img/orchestrator_landing.png
new file mode 100644
index 0000000..4770f40
Binary files /dev/null and b/collections/documentation/dashboard/img/orchestrator_landing.png differ
diff --git a/collections/documentation/dashboard/img/owncloud_create_user.png b/collections/documentation/dashboard/img/owncloud_create_user.png
new file mode 100644
index 0000000..ba0badc
Binary files /dev/null and b/collections/documentation/dashboard/img/owncloud_create_user.png differ
diff --git a/collections/documentation/dashboard/img/owncloud_credentials.png b/collections/documentation/dashboard/img/owncloud_credentials.png
new file mode 100644
index 0000000..8f1ad29
Binary files /dev/null and b/collections/documentation/dashboard/img/owncloud_credentials.png differ
diff --git a/collections/documentation/dashboard/img/owncloud_details.png b/collections/documentation/dashboard/img/owncloud_details.png
new file mode 100644
index 0000000..efe6136
Binary files /dev/null and b/collections/documentation/dashboard/img/owncloud_details.png differ
diff --git a/collections/documentation/dashboard/img/owncloud_enabled.png b/collections/documentation/dashboard/img/owncloud_enabled.png
new file mode 100644
index 0000000..8877f6b
Binary files /dev/null and b/collections/documentation/dashboard/img/owncloud_enabled.png differ
diff --git a/collections/documentation/dashboard/img/owncloud_logo.svg b/collections/documentation/dashboard/img/owncloud_logo.svg
new file mode 100644
index 0000000..4959a9f
--- /dev/null
+++ b/collections/documentation/dashboard/img/owncloud_logo.svg
@@ -0,0 +1,159 @@
+
+
+
+
+
+ image/svg+xml
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
diff --git a/collections/documentation/dashboard/img/owncloud_logout.png b/collections/documentation/dashboard/img/owncloud_logout.png
new file mode 100644
index 0000000..ba21a9f
Binary files /dev/null and b/collections/documentation/dashboard/img/owncloud_logout.png differ
diff --git a/collections/documentation/dashboard/img/owncloud_show.png b/collections/documentation/dashboard/img/owncloud_show.png
new file mode 100644
index 0000000..6b16ee3
Binary files /dev/null and b/collections/documentation/dashboard/img/owncloud_show.png differ
diff --git a/collections/documentation/dashboard/img/owncloud_tfconnect.png b/collections/documentation/dashboard/img/owncloud_tfconnect.png
new file mode 100644
index 0000000..d386991
Binary files /dev/null and b/collections/documentation/dashboard/img/owncloud_tfconnect.png differ
diff --git a/collections/documentation/dashboard/img/owncloud_users.png b/collections/documentation/dashboard/img/owncloud_users.png
new file mode 100644
index 0000000..aec7c03
Binary files /dev/null and b/collections/documentation/dashboard/img/owncloud_users.png differ
diff --git a/collections/documentation/dashboard/img/owncloud_visit.png b/collections/documentation/dashboard/img/owncloud_visit.png
new file mode 100644
index 0000000..5624e89
Binary files /dev/null and b/collections/documentation/dashboard/img/owncloud_visit.png differ
diff --git a/collections/documentation/dashboard/img/polkadot_extension_.png b/collections/documentation/dashboard/img/polkadot_extension_.png
new file mode 100644
index 0000000..bf3f6fd
Binary files /dev/null and b/collections/documentation/dashboard/img/polkadot_extension_.png differ
diff --git a/collections/documentation/dashboard/img/pricing_calculator.png b/collections/documentation/dashboard/img/pricing_calculator.png
new file mode 100644
index 0000000..31f21eb
Binary files /dev/null and b/collections/documentation/dashboard/img/pricing_calculator.png differ
diff --git a/collections/documentation/dashboard/img/pro_manager5.png b/collections/documentation/dashboard/img/pro_manager5.png
new file mode 100644
index 0000000..791dae0
Binary files /dev/null and b/collections/documentation/dashboard/img/pro_manager5.png differ
diff --git a/collections/documentation/dashboard/img/pro_manager6.png b/collections/documentation/dashboard/img/pro_manager6.png
new file mode 100644
index 0000000..2b77950
Binary files /dev/null and b/collections/documentation/dashboard/img/pro_manager6.png differ
diff --git a/collections/documentation/dashboard/img/profile_manager1.png b/collections/documentation/dashboard/img/profile_manager1.png
new file mode 100644
index 0000000..73d59d5
Binary files /dev/null and b/collections/documentation/dashboard/img/profile_manager1.png differ
diff --git a/collections/documentation/dashboard/img/profile_manager2.png b/collections/documentation/dashboard/img/profile_manager2.png
new file mode 100644
index 0000000..98f2154
Binary files /dev/null and b/collections/documentation/dashboard/img/profile_manager2.png differ
diff --git a/collections/documentation/dashboard/img/profile_manager3.png b/collections/documentation/dashboard/img/profile_manager3.png
new file mode 100644
index 0000000..407780c
Binary files /dev/null and b/collections/documentation/dashboard/img/profile_manager3.png differ
diff --git a/collections/documentation/dashboard/img/rent_contracts.png b/collections/documentation/dashboard/img/rent_contracts.png
new file mode 100644
index 0000000..170ca57
Binary files /dev/null and b/collections/documentation/dashboard/img/rent_contracts.png differ
diff --git a/collections/documentation/dashboard/img/select_to_delete_domain.png b/collections/documentation/dashboard/img/select_to_delete_domain.png
new file mode 100644
index 0000000..1ae8ce8
Binary files /dev/null and b/collections/documentation/dashboard/img/select_to_delete_domain.png differ
diff --git a/collections/documentation/dashboard/img/sidebar_1 (copy)_full.png b/collections/documentation/dashboard/img/sidebar_1 (copy)_full.png
new file mode 100644
index 0000000..9e01af9
Binary files /dev/null and b/collections/documentation/dashboard/img/sidebar_1 (copy)_full.png differ
diff --git a/collections/documentation/dashboard/img/sidebar_1.png b/collections/documentation/dashboard/img/sidebar_1.png
new file mode 100644
index 0000000..4779b4c
Binary files /dev/null and b/collections/documentation/dashboard/img/sidebar_1.png differ
diff --git a/collections/documentation/dashboard/img/sidebar_2 (copy)_full.png b/collections/documentation/dashboard/img/sidebar_2 (copy)_full.png
new file mode 100644
index 0000000..f7d8600
Binary files /dev/null and b/collections/documentation/dashboard/img/sidebar_2 (copy)_full.png differ
diff --git a/collections/documentation/dashboard/img/sidebar_2.png b/collections/documentation/dashboard/img/sidebar_2.png
new file mode 100644
index 0000000..58e913b
Binary files /dev/null and b/collections/documentation/dashboard/img/sidebar_2.png differ
diff --git a/collections/documentation/dashboard/img/sidebar_3 (copy)_full.png b/collections/documentation/dashboard/img/sidebar_3 (copy)_full.png
new file mode 100644
index 0000000..8924f42
Binary files /dev/null and b/collections/documentation/dashboard/img/sidebar_3 (copy)_full.png differ
diff --git a/collections/documentation/dashboard/img/sidebar_3.png b/collections/documentation/dashboard/img/sidebar_3.png
new file mode 100644
index 0000000..565715d
Binary files /dev/null and b/collections/documentation/dashboard/img/sidebar_3.png differ
diff --git a/collections/documentation/dashboard/img/sidebar_4 (copy)_full.png b/collections/documentation/dashboard/img/sidebar_4 (copy)_full.png
new file mode 100644
index 0000000..7a7e165
Binary files /dev/null and b/collections/documentation/dashboard/img/sidebar_4 (copy)_full.png differ
diff --git a/collections/documentation/dashboard/img/sidebar_4.png b/collections/documentation/dashboard/img/sidebar_4.png
new file mode 100644
index 0000000..8f94e9e
Binary files /dev/null and b/collections/documentation/dashboard/img/sidebar_4.png differ
diff --git a/collections/documentation/dashboard/img/simulator.png b/collections/documentation/dashboard/img/simulator.png
new file mode 100644
index 0000000..b85a528
Binary files /dev/null and b/collections/documentation/dashboard/img/simulator.png differ
diff --git a/collections/documentation/dashboard/img/statistics.png b/collections/documentation/dashboard/img/statistics.png
new file mode 100644
index 0000000..a216231
Binary files /dev/null and b/collections/documentation/dashboard/img/statistics.png differ
diff --git a/collections/documentation/dashboard/img/twin.png b/collections/documentation/dashboard/img/twin.png
new file mode 100644
index 0000000..1e62185
Binary files /dev/null and b/collections/documentation/dashboard/img/twin.png differ
diff --git a/collections/documentation/dashboard/img/vm_landing.png b/collections/documentation/dashboard/img/vm_landing.png
new file mode 100644
index 0000000..21211a3
Binary files /dev/null and b/collections/documentation/dashboard/img/vm_landing.png differ
diff --git a/collections/documentation/dashboard/play_go.md b/collections/documentation/dashboard/play_go.md
new file mode 100644
index 0000000..5bb8332
--- /dev/null
+++ b/collections/documentation/dashboard/play_go.md
@@ -0,0 +1,5 @@
+- Choose one of the networks:
+ - https://dashboard.dev.grid.tf for Devnet.
+ - https://dashboard.qa.grid.tf for QAnet.
+ - https://dashboard.test.grid.tf for Testnet.
+ - https://dashboard.grid.tf for Mainnet.
\ No newline at end of file
diff --git a/collections/documentation/dashboard/solutions/add_domain.md b/collections/documentation/dashboard/solutions/add_domain.md
new file mode 100644
index 0000000..6883c97
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/add_domain.md
@@ -0,0 +1,107 @@
+ Add a Domain to a VM
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Preparation](#preparation)
+- [Add New Domain](#add-new-domain)
+- [Domains List](#domains-list)
+- [Delete a Domain](#delete-a-domain)
+- [Questions and Feedback](#questions-and-feedback)
+
+***
+
+## Introduction
+
+We cover the overall process to add a domain to a virtual machine running on the ThreeFold Grid.
+
+## Preparation
+
+- Deploy a virtual machine
+- Click on the button **Manage Domains** under **Actions**
+
+![](../img/add_domain_6.png)
+
+- Open the **Add New Domain** tab
+
+![](../img/add_domain_10.png)
+
+## Add New Domain
+
+We cover the different domain parameters presented in the **Add New Domain** tab.
+
+- **Subdomain**
+ - The subdomain is used to reference to the complete domain name. It is randomly generated, but the user can write a specific subdomain name.
+ - The subdomain prefix (e.g. **fvm3748domainguide**) is decided as follows:
+ - Solution name (e.g. **fvm**)
+ - Twin ID (e.g. **3748**)
+ - Deployment name (e.g. **domainguide**)
+ - The complete subdomain is thus composed of the subdomain prefix mentioned above and the subdomain entered in the **Subdomain** field.
+- **Custom domain name**
+ - You can also use a custom domain.
+ - In this case, instead of having a gateway subdomain and a gateway name as your domain, the domain will be the custom domain entered in this field.
+ - If you select **Custom domain**, make sure to set a DNS A record pointing to the gateway IP address on your domain name registrar.
+
+![Custom Domain Name](../img/add_domain_8.png)
+
+- **Select domain**
+ - Choose a gateway for your domain.
+
+- **Port**
+ - Choose the port that exposes your application instance on the virtual machine which the domain will point to.
+ - By default, it is set to **80**.
+
+- **TLS Passthrough**
+ - Disabling TLS passthrough will let the gateway terminate the traffic.
+ - Enabling TLS passthrough will let the backend service terminate the traffic.
+
+- **Network Name**
+ - This is the name of the WireGuard interface network (read-only field).
+
+- **IP Address**
+ - This is the WireGuard IP address (read-only field).
+
+Once you've filled the domain parameters, click on the **Add** button. The message **Successfully deployed gateway** will be presented once the domain is properly added.
+
+![Success Domain](../img/add_new_domain_success.png)
+
+## Domains List
+
+Once your domain is set, you can access the **Domains List** tab to consult its parameters. To visit the domain, simply click on the **Visit** button under **Actions**.
+
+![List Domain For VM](../img/add_domain_9.png)
+
+* **Name**
+ * The name is the subdomain (without the prefix)
+* **Contract ID**
+ * Contract ID of the domain
+* **Domain**
+ * Without a custom domain (default)
+ * The complete domain name (e.g. `fvm3748domainguidextebgpt.gent01.dev.grid.tf`) is composed of the subdomain prefix, the subdomain and the gateway domain.
+ - The subdomain prefix (e.g. `fvm3748domainguide`), as mentioned above.
+ - The subdomain (e.g. `xtebgpt`), chosen during the domain creation.
+ - The gateway domain (e.g. `gent01.dev.grid.tf`), based on the chosen gateway.
+ - With a custom domain
+ - The domain will be your custom domain (`e.g. threefold.pro`).
+* **TLS Passthrough**
+ * The TLS passthrough status can be either **Yes** or **No**.
+* **Backend**
+ * The WireGuard IP and the chosen port of the domain (e.g. `http://10.20.4.2:80`).
+* **Status**
+ * **OK** is displayed when the domain is properly set.
+* **Actions**
+ * Use the **Visit** button to open the domain URL.
+
+At all time, you can click on **Reload** to reload the Domains List parameters.
+
+## Delete a Domain
+
+To delete a domain, open the **Manage Domains** window, in the tab **Domains lists** select the domain you wish to delete and click **Delete**.
+
+![Select To Delete Domain](../img/add_domain_11.png)
+
+By clicking the **Delete** button, the deletion will start and the domain will be deleted from this virtual machine.
+
+## Questions and Feedback
+
+If you have any questions, you can ask the ThreeFold community for help on the [ThreeFold Forum](http://forum.threefold.io/) or on the [ThreeFold Grid Tester Community](https://t.me/threefoldtesting) on Telegram.
\ No newline at end of file
diff --git a/collections/documentation/dashboard/solutions/algorand.md b/collections/documentation/dashboard/solutions/algorand.md
new file mode 100644
index 0000000..21c25c7
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/algorand.md
@@ -0,0 +1,97 @@
+ Algorand
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Prerequisites](#prerequisites)
+ - [Algorand Structure](#algorand-structure)
+- [Run Default Node](#run-default-node)
+- [Run Relay Node](#run-relay-node)
+- [Run Participant Node](#run-participant-node)
+- [Run Indexer Node](#run-indexer-node)
+- [Select Capacity](#select-capacity)
+
+***
+
+## Introduction
+
+[Algorand](https://www.algorand.com/) builds technology that accelerates the convergence between decentralized and traditional finance by enabling the simple creation of next-generation financial products, protocols, and exchange of value.
+
+## Prerequisites
+
+- Make sure you have a [wallet](../wallet_connector.md)
+- From the sidebar click on **Applications**
+- Click on **Algorand**
+
+### Algorand Structure
+
+- Algorand has two main [types](https://developer.algorand.org/docs/run-a-node/setup/types/#:~:text=The%20Algorand%20network%20is%20comprised,%2C%20and%20non%2Drelay%20nodes.) of nodes (Relay or Participant) you can run also a 4 networks you can run your node against. Combining the types you can get:
+ - Defualt:
+ This is a Non-relay and Non-participant
+ It can run on (Devnet, Testnet, Betanet, Mainnet)
+ - Relay:
+ Relay node Can't be participant.
+ It can run only on (Testnet, Mainnet)
+ - Participant:
+ Can run on any of the four nets.
+ - Indexer:
+ It is a default node but with Archival Mode enbled which will make you able to query the data of the blockchain.
+
+## Run Default Node
+
+The basic type. you select any network you want. and for the node type select Default.
+![defaultdep](./img/solutions_algorand.png)
+
+after the deployment is done. `ssh` to the node and run `goal node status`
+![defaulttest](./img/algorand_defaulttest.png)
+here you see your node run against mainnet.
+
+## Run Relay Node
+
+Relay nodes are where other nodes connect. Therefore, a relay node must be able to support a large number of connections and handle the processing load associated with all the data flowing to and from these connections. Thus, relay nodes require significantly more power than non-relay nodes. Relay nodes are always configured in archival mode.
+
+The relay node must be publicaly accessable. so it must have public ip.
+![relaydep](./img/algorand_relaydep.png)
+
+after the deployment is done. `ssh` to the node and run `goal node status` to see the status of the node. and also you can check if the right port is listening (:4161 for testnet, and :4160 for mainnet)
+![relaytest](./img/algorand_relaytest.png)
+
+The next step accourding to the [docs](https://developer.algorand.org/docs/run-a-node/setup/types/#relay-node) is to register your `ip:port` on Algorand Public SRV.
+
+## Run Participant Node
+
+Participation means participation in the Algorand consensus protocol. An account that participates in the Algorand consensus protocol is eligible and available to be selected to propose and vote on new blocks in the Algorand blockchain.
+Participation node is responsible for hosting participation keys for one or more online accounts.
+
+What you need?
+- Account mnemonics on the network you deploy on (offline) you can check the status for you account on the AlgoExplorer. search by your account id.
+
+ The account needs to have some microAlgo to sign the participation transaction.
+ - [Main net explorer](https://algoexplorer.io/)
+ - [Test net explorer](https://testnet.algoexplorer.io/)
+
+- First Round: is the first block you need your participaiton node to validate from. you can choose the last block form the explorer.
+ ![partexp](./img/algorand_partexp.png)
+- Last Round: is the final block your node can validate. let's make it 30M
+
+![partdep](./img/algorand_partdep.png)
+
+after the deployment is done. `ssh` to the node and run `goal node status` to see the status of the node. you see it do catchup. and the fast catchup is to make the node sync with the latest block faster by only fetch the last 1k blocks. after it done it will start create the participation keys.
+![partstatus](./img/algorand_partstatus.png)
+
+now if you check the explorer you can see the status of the account turned to Online
+![partonl](./img/algorand_partonl.png)
+
+## Run Indexer Node
+
+The primary purpose of this Indexer is to provide a REST API interface of API calls to support searching the Algorand Blockchain. The Indexer REST APIs retrieve the blockchain data from a PostgreSQL compatible database that must be populated. This database is populated using the same indexer instance or a separate instance of the indexer which must connect to the algod process of a running Algorand node to read block data. This node must also be an Archival node to make searching the entire blockchain possible.
+
+![indexernode](./img/algorand_indexernode.png)
+
+After it finish you can access the indexer API at port `8980` and here are the [endpoint](https://developer.algorand.org/docs/rest-apis/indexer/) you can access.
+
+## Select Capacity
+
+The default scinario the capacity is computed based on the node (network/type) accourding to this [reference](https://howbigisalgorand.com/).
+But you still can change this only to higher values by selecting the option `Set Custom Capacity`
+
\ No newline at end of file
diff --git a/collections/documentation/dashboard/solutions/basic_environments_readme.md b/collections/documentation/dashboard/solutions/basic_environments_readme.md
new file mode 100644
index 0000000..d1bd532
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/basic_environments_readme.md
@@ -0,0 +1,11 @@
+ Basic Environments
+
+ Table of Contents
+
+- [Virtual Machines](./vm_intro.md)
+ - [Micro and Full VM Differences ](./vm_differences.md)
+ - [Full Virtual Machine](./fullVm.md)
+ - [Micro Virtual Machine](./vm.md)
+- [Kubernetes](./k8s.md)
+- [NixOS MicroVM](./nixos_micro.md)
+- [Add a Domain](./add_domain.md)
\ No newline at end of file
diff --git a/collections/documentation/dashboard/solutions/caprover.md b/collections/documentation/dashboard/solutions/caprover.md
new file mode 100644
index 0000000..32034d7
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/caprover.md
@@ -0,0 +1,165 @@
+ CapRover
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Requirements](#requirements)
+- [Configs Tab](#configs-tab)
+- [Admin and Workers Tabs](#admin-and-workers-tabs)
+- [The Domain Name](#the-domain-name)
+ - [Domain Name Example](#domain-name-example)
+- [How to Know the IP Address](#how-to-know-the-ip-address)
+- [How to Access the Admin Interface](#how-to-access-the-admin-interface)
+- [How to Work with CapRover](#how-to-work-with-caprover)
+
+***
+
+## Introduction
+
+CapRover is an extremely easy to use app/database deployment & web server manager for your NodeJS, Python, PHP, ASP.NET, Ruby, MySQL, MongoDB, Postgres, WordPress (and etc...) applications!
+
+It's blazingly fast and very robust as it uses Docker, nginx, LetsEncrypt and NetData under the hood behind its simple-to-use interface.
+
+- CLI for automation and scripting
+- Web GUI for ease of access and convenience
+- No lock-in! Remove CapRover and your apps keep working!
+- Docker Swarm under the hood for containerization and clustering
+- Nginx (fully customizable template) under the hood for load-balancing
+- Let's Encrypt under the hood for free SSL (HTTPS)
+
+Caprover is a very cool management app for containers based on Docker Swarm.
+
+It has following benefits :
+
+- easy to deploy apps (in seconds)
+- easy to create new apps
+- super good monitoring
+- can be extended over the TFGrid
+
+## Requirements
+
+- Make sure you have a [wallet](../wallet_connector.md)
+- From the sidebar click on **Orchestrators**
+- Click on **CapRover**
+
+## Configs Tab
+
+![ ](./img/solutions_caprover.png)
+
+- Enter domain for you Caprover instance, Be very careful about the domain name: it needs to be a wildcard domain name you can configure in your chosen domain name system.
+- Enter password for you Caprover instance.
+
+## Admin and Workers Tabs
+
+![ ](./img/solutions_caprover_leader.png)
+
+![ ](./img/solutions_caprover_workers.png)
+Note: Worker nodes only accept SSH keys of RSA format.
+
+Deployment will take couple of minutes.
+
+## The Domain Name
+
+As per the [CapRover documentation](https://caprover.com/docs/get-started.html), you need to point a wildcard DNS entry to the VM IP address of your CapRover instance. You have to do this after having deployed the CapRover instance, otherwise you won't have access to the VM IP address.
+
+Let’s say your domain is **example.com** and your subdomain is **subdomain**. You can set **\*.subdomain.example.com** as an A record in your DNS settings to point to the VM IP address of the server hosting the CapRover instance, where **\*** acts as the wildcard. To do this, go to the DNS settings of your domain name registrar, and set a wild card A record entry.
+
+On your domain name registrar, you can manage your DNS settings as such, with **subdomain** as an example:
+
+| Record | Host | Value | TTL |
+| ------ | ------------- | ------------- | --------- |
+| A | @ | VM IP address | Automatic |
+| A | subdomain | VM IP address | Automatic |
+| A | \*.subdomain | VM IP address | Automatic |
+
+We note here that **@** is the root domain (@ takes the value of your domain name, e.g. **example** in **example.com**), **subdomain** is the name of your subdomain (it can be anything you want), and **\*.subdomain** is the wildcard for **subdomain**. If you don't want to use a subdomain, but only the domain, you could use a wildcard linked to the domain instead of the subdomain (e.g. put **\*** instead of **\*.subdomain** in the column **Host**).
+
+Once you've point a wildcard DNS entry to your CapRover IP address and that the DNS is properly propagated, you can click the **Admin Panel** button to access CapRover. This will lead you to the following URL (with **subdomain.example.com** as an example):
+
+> captain.subdomain.example.com
+
+Note that, to confirm the DNS propagation, you can use a [DNS lookup tool](https://mxtoolbox.com/DNSLookup.aspx). As an example, you can use the URL **captain.subdomain.example.com** to check if the IP address resolves to the VM IP address.
+
+### Domain Name Example
+
+In the following example, we pick ```apps.openly.life``` which is a domain name that will point to the IP address of the CapRover instance (which we only know after deployment).
+
+![ ](./img/domain_name_caprover_config.png)
+
+> Note how the *.apps.openly.life points to the public IPv4 address that has been returned from the deployment.
+
+## How to Know the IP Address
+
+Go back to your CapRover weblet and go to the deployment list. Click on `Show Details`.
+
+![ ](./img/solution_caprover_list.png)
+
+- The public IPv4 address is visible in here
+- Now you can configure the domain name (see above, don't forget to point the wildcard domain to the public IP address)
+
+Click on details if you want to see more details
+
+```json
+
+{
+ "version": 0,
+ "name": "caprover_leader_cr_156e44f0",
+ "created": 1637843368,
+ "status": "ok",
+ "message": "",
+ "flist": "https://hub.grid.tf/samehabouelsaad.3bot/tf-caprover-main-a4f186da8d.flist",
+ "publicIP": {
+ "ip": "185.206.122.136/24",
+ "gateway": "185.206.122.1"
+ },
+ "planetary": false,
+ "yggIP": "",
+ "interfaces": [
+ {
+ "network": "caprover_network_cr_156e44f0",
+ "ip": "10.200.4.2"
+ }
+ ],
+ "capacity": {
+ "cpu": 4,
+ "memory": 8192
+ },
+ "mounts": [
+ {
+ "name": "data0",
+ "mountPoint": "/var/lib/docker",
+ "size": 107374182400,
+ "state": "ok",
+ "message": ""
+ }
+ ],
+ "env": {
+ "SWM_NODE_MODE": "leader",
+ "CAPROVER_ROOT_DOMAIN": "apps.openly.life",
+ "PUBLIC_KEY": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQC/9RNGKRjHvViunSOXhBF7EumrWvmqAAVJSrfGdLaVasgaYK6tkTRDzpZNplh3Tk1aowneXnZffygzIIZ82FWQYBo04IBWwFDOsCawjVbuAfcd9ZslYEYB3QnxV6ogQ4rvXnJ7IHgm3E3SZvt2l45WIyFn6ZKuFifK1aXhZkxHIPf31q68R2idJ764EsfqXfaf3q8H3u4G0NjfWmdPm9nwf/RJDZO+KYFLQ9wXeqRn6u/mRx+u7UD+Uo0xgjRQk1m8V+KuLAmqAosFdlAq0pBO8lEBpSebYdvRWxpM0QSdNrYQcMLVRX7IehizyTt+5sYYbp6f11WWcxLx0QDsUZ/J"
+ },
+ "entrypoint": "/sbin/zinit init",
+ "metadata": "",
+ "description": "caprover leader machine/node"
+}
+```
+
+## How to Access the Admin Interface
+
+Make sure that you've point a wildcard DNS entry to your CapRover IP address (e.g. **185.206.122.136** in our example), as explained [here](#the-domain-name).
+
+* To access the CapRover admin interface, you can click the **Admin Panel** button or you can use the following admin URL template: **https://captain.subdomain.example.com**.
+ * Note the prefix **captain** and the usage of our wildcard domain.
+
+* The admin password is generated and visible behind the `Show Details` button of your CapRover deployment.
+
+![ ](./img/caprover_login.png)
+
+* You should now see the following screen:
+
+![ ](./img/captain_login+weblet_caprover_.png)
+
+## How to Work with CapRover
+
+* [CapRover Admin Tutorial](./caprover_admin.md)
+* [CapRover Worker Tutorial](./caprover_worker.md)
diff --git a/collections/documentation/dashboard/solutions/caprover_admin.md b/collections/documentation/dashboard/solutions/caprover_admin.md
new file mode 100644
index 0000000..47dd8fd
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/caprover_admin.md
@@ -0,0 +1,59 @@
+ CapRover Admin
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Step 1: Enable HTTPS](#step-1-enable-https)
+- [Step 2: Add a Default Docker Registry](#step-2-add-a-default-docker-registry)
+- [Step 3: Deploy an App](#step-3-deploy-an-app)
+- [Step 4: Enable Monitoring](#step-4-enable-monitoring)
+- [Step 5: Change Your Password](#step-5-change-your-password)
+
+***
+
+## Introduction
+
+We present the steps to manage a CapRover Admin node.
+
+## Step 1: Enable HTTPS
+
+![ ](./img/enable_https_caprover.png)
+
+You need to specify your email address.
+
+You will have to login again.
+
+![ ](./img/caprover_https_activated.png)
+
+> Now force https.
+
+You will have to login again, and you should notice https is now used.
+
+## Step 2: Add a Default Docker Registry
+
+You'll have to add a default docker registry so other CapRover nodes in the cluster can download images from, and it can be self-hosted (managed by CapRover itself), to add it, go to `Cluster` -> `Docker Registry Configuration`.
+
+![ ](./img/caprover_docker_registry.png)
+
+You can check [official documentation](https://caprover.com/docs/app-scaling-and-cluster.html#setup-docker-registry) to know more about Docker registry options.
+
+## Step 3: Deploy an App
+
+![ ](./img/deploy_app_caprover1.png)
+
+just go to apps & follow the instructions, there is much more info on caprover website.
+
+## Step 4: Enable Monitoring
+
+![ ](./img/caprover_monitoring_start_.png)
+
+You should now see
+
+![ ](./img/caprover_monitoring_2_.png)
+
+## Step 5: Change Your Password
+
+- Go to `Settings` and change your password. This is important for your own security.
+
+
+> Further information regarding the process of attaching a new node to the cluster can be found through the following documentation link: [Attach a New Node to the Cluster](./caprover_worker.md/#step-2-attach-a-new-node-to-the-cluster)
\ No newline at end of file
diff --git a/collections/documentation/dashboard/solutions/caprover_worker.md b/collections/documentation/dashboard/solutions/caprover_worker.md
new file mode 100644
index 0000000..5269e23
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/caprover_worker.md
@@ -0,0 +1,42 @@
+ CapRover Worker
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Step 1: Add a Default Docker Registry](#step-1-add-a-default-docker-registry)
+- [Step 2: Attach a New Node to the Cluster](#step-2-attach-a-new-node-to-the-cluster)
+
+***
+
+## Introduction
+
+We present the steps to manage a CapRover Worker node.
+
+## Step 1: Add a Default Docker Registry
+
+You'll have to add a default docker registry so other CapRover nodes in the cluster can download images from, and it can be self-hosted (managed by CapRover itself), to add it, go to `Cluster` -> `Docker Registry Configuration`.
+
+![ ](./img/caprover_docker_registry.png)
+
+- Click `Add Self-Hosted Registry` button, then click `Enable Self-Hosted Registry`
+
+![ ](./img/caprover_docker_default_registry.png)
+
+You can check [official documentation](https://caprover.com/docs/app-scaling-and-cluster.html#setup-docker-registry) to know more about Docker registry options.
+
+
+
+## Step 2: Attach a New Node to the Cluster
+
+![ ](./img/caprover_add_worker.png)
+
+- Add the public IPv4 address that has been returned from the worker deployment in the `New node IP Address` field.
+- Add your `SSH private key` (you can use this command `cat ~/.ssh/id_rsa` to get your private key).
+- Click `Join cluster` button.
+
+You should see the new added node under **Current Cluster Nodes**
+![ ](./img/caprover_node_added.png)
+
+If you faced any problem you can use the `Alternative method`.
+
+Also you can check for Troubleshooting instruction on [Caprover Troubleshooting](https://caprover.com/docs/troubleshooting.html#second)
\ No newline at end of file
diff --git a/collections/documentation/dashboard/solutions/casper.md b/collections/documentation/dashboard/solutions/casper.md
new file mode 100644
index 0000000..5071c09
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/casper.md
@@ -0,0 +1,51 @@
+ CasperLabs
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Deployment](#deployment)
+
+***
+
+## Introduction
+
+[Casper Network](https://casperlabs.io/) is a blockchain protocol built from the ground up to remain true to core Web3 principles and adapt to the needs of our evolving world.
+
+- Make sure you have a [wallet](../wallet_connector.md)
+- From the sidebar click on **Applications**
+- Click on **Casperlabs**
+
+## Deployment
+
+__Process__ :
+
+![ ](./img/solutions_casperlabs.png)
+
+- Enter an Application Name. It's used in generating a unique subdomain on one of the gateways on the network alongside your twin ID. Ex. ***cl98casp*.gent02.dev.grid.tf**
+
+- Select a capacity package:
+ - **Small**: {cpu: 2, memory: 4, diskSize: 100 }
+ - **Medium**: {cpu: 4, memory: 16, diskSize: 500 }
+ - **Large**: {cpu: 8, memory: 32, diskSize: 100 }
+ - Or choose a **Custom** plan
+- Choose the network
+ - `Public IPv4` flag gives the virtual machine a Public IPv4
+
+- `Dedicated` flag to retrieve only dedeicated nodes
+- `Certified` flag to retrieve only certified nodes
+- Choose the location of the node
+ - `Region`
+ - `Country`
+ - `Farm Name`
+- Choose the node to deploy on
+> Or you can select a specific node with manual selection.
+- `Custom Domain` flag lets the user to use a custom domain
+- Choose a gateway node to deploy your Casperlab instance on.
+
+After that is done you can see a list of all of your deployed instances
+
+![ ](./img/casper4.png)
+
+Click on ***Visit*** to go to the homepage of your Casperlabs instance! The node takes a long time in order for the RPC service to be ready so be patient!
+
+![ ](./img/casper5.png)
\ No newline at end of file
diff --git a/collections/documentation/dashboard/solutions/discourse.md b/collections/documentation/dashboard/solutions/discourse.md
new file mode 100644
index 0000000..9bdff3c
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/discourse.md
@@ -0,0 +1,53 @@
+ Discourse
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Deployment](#deployment)
+
+***
+
+## Introduction
+
+[Discourse](https://www.discourse.org/) is the 100% open source discussion platform built for the next decade of the Internet. Use it as a mailing list, discussion forum, long-form chat room, and more!
+
+- Make sure you have a [wallet](../wallet_connector.md)
+- From the sidebar click on **Applications**
+- Click on **Discourse**
+
+## Deployment
+
+![ ](./img/solutions_discourse.png)
+
+- Enter an Application Name. It's used in generating a unique subdomain on one of the gateways on the network alongside your twin ID. Ex. ***dc98newdisc*.gent02.dev.grid.tf**
+
+- Enter administrator information including **Email**. This admin will have full permission on the deployed instance.
+- Select a capacity package:
+ - **Small**: {cpu: 1, memory: 2, diskSize: 15 }
+ - **Medium**: {cpu: 2, memory: 4, diskSize: 50 }
+ - **Large**: {cpu: 4, memory: 16, diskSize: 100 }
+ - Or choose a **Custom** plan
+
+- `Dedicated` flag to retrieve only dedeicated nodes
+- `Certified` flag to retrieve only certified nodes
+- Choose the location of the node
+ - `Region`
+ - `Country`
+ - `Farm Name`
+
+- Choose the node to deploy on
+> Or you can select a specific node with manual selection.
+- `Custom Domain` flag lets the user to use a custom domain
+- Choose a gateway node to deploy your Discourse instance on.
+
+Unlike other solutions, Discourse requires that you have an SMTP server. So make sure you fill the fields in the **Mail Server** tab in order to deploy your instance successfully.
+
+![ ](./img/discourse4.png)
+
+After that is done you can see a list of all of your deployed instances
+
+![ ](./img/discourse5.png)
+
+Click on ***Visit*** to go to the homepage of your Discourse instance!
+
+![ ](./img/discourse6.png)
diff --git a/collections/documentation/dashboard/solutions/fullVm.md b/collections/documentation/dashboard/solutions/fullVm.md
new file mode 100644
index 0000000..59b01e6
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/fullVm.md
@@ -0,0 +1,108 @@
+ Full Virtual Machine
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Deployment](#deployment)
+- [Difference Between Full VM and Micro VM](#difference-between-full-vm-and-micro-vm)
+- [Manually Mounting Additional Disk](#manually-mounting-additional-disk)
+ - [Check All Disks Attached to the VM](#check-all-disks-attached-to-the-vm)
+ - [Create a Mount Directory](#create-a-mount-directory)
+ - [New file system](#new-file-system)
+ - [Mount drive](#mount-drive)
+
+***
+
+## Introduction
+
+We present the steps to deploy a full VM on the TFGrid.
+
+## Deployment
+
+Deploy a new full virtual machine on the Threefold Grid
+
+- Make sure you have a [wallet](../wallet_connector.md)
+- From the sidebar click on **Applications**
+- Click on **Full Virtual Machine**
+
+**Process:**
+
+![ ](./img/solutions_fullvm.png)
+
+- Fill in the instance name: it's used to reference the Full VM in the future.
+- Choose the image from the drop down (e.g Alpine, Ubuntu) or you can click on `Other` and manually specify the flist URL and the entrypoint.
+- Select a capacity package:
+ - **Small**: {cpu: 1, memory: 2, diskSize: 25 }
+ - **Medium**: {cpu: 2, memory: 4, diskSize: 50 }
+ - **Large**: {cpu: 4, memory: 16, diskSize: 100}
+ - Or choose a **Custom** plan
+- Choose the network
+ - `Public IPv4` flag gives the virtual machine a Public IPv4
+ - `Public IPv6` flag gives the virtual machine a Public IPv6
+ - `Planetary Network` to connect the Virtual Machine to Planetary network
+ - `Myceluim` to enable mycelium on the virtual machine
+ - `Wireguard Access` to add a wireguard access to the Virtual Machine
+- `GPU` flag to add GPU to the Virtual machine
+ - To deploy a Full VM with GPU, you first need to [rent a dedicated node](../../dashboard/deploy/dedicated_machines.md)
+- `Dedicated` flag to retrieve only dedicated nodes
+- `Certified` flag to retrieve only certified nodes
+- Choose the location of the node
+ - `Country`
+ - `Farm Name`
+- Choose the node to deploy the Full Virtual Machine on
+ ![](./img/node_selection.png)
+
+You can attach one or more disks to the Virtual Machine by clicking on the Disks tab and the plus `+` sign and specify the following parameters
+![ ](./img/new_vm3.png)
+
+- Disk name
+- Disk size
+
+in the bottom of the page you can see a list of all of the virtual machines you deployed. you can click on `Show details` for more details
+
+![ ](./img/new_vm5.png)
+You can also go to JSON tab for full details
+![ ](./img/new_vm6.png)
+
+## Difference Between Full VM and Micro VM
+
+- Full VM contains a default disk attached to it which is not the case in the Micro VM where you needed to make sure to attach a disk to it or the VM will fail
+- The default disk is mounted on / so if you want to attach any additional disks, you have to choose a different mounting point
+- Only cloud init flists can be deployed on Full VM. You can check official Threefold flists [here](https://hub.grid.tf/tf-official-vms)
+- In Full VM, you need to mount the additional disks manually after the VM is deployed
+
+## Manually Mounting Additional Disk
+
+- You can follow the following commands to add your disk manually:
+
+### Check All Disks Attached to the VM
+
+```bash
+fdisk -l
+```
+
+The additional disk won't be mounted and you won't find it listed
+
+```bash
+df -h
+```
+
+### Create a Mount Directory
+
+```bash
+sudo mkdir /hdd6T
+```
+
+### New file system
+
+```bash
+sudo mkfs.ext4 /dev/vdb
+```
+
+### Mount drive
+
+```bash
+sudo mount /dev/vdb /hdd6T/
+```
+
+![mounting additional disk](./img/fullvm6.png)
diff --git a/collections/documentation/dashboard/solutions/funkwhale.md b/collections/documentation/dashboard/solutions/funkwhale.md
new file mode 100644
index 0000000..8940728
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/funkwhale.md
@@ -0,0 +1,59 @@
+ Funkwhale
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Prerequisites](#prerequisites)
+- [Deployment](#deployment)
+
+***
+
+## Introduction
+
+[Funkwhale](https://funkwhale.audio/) is social platform to enjoy and share music.
+Funkwhale is a community-driven project that lets you listen and share music and audio within a decentralized, open network.
+
+## Prerequisites
+
+- Make sure you have a [wallet](../wallet_connector.md)
+- From the sidebar click on **Applications**
+- Click on **Funkwhale**
+
+## Deployment
+
+__Process__ :
+
+![ ](./img/solutions_funkwhale.png)
+
+- Enter an Application Name. It's used in generating a unique subdomain on one of the gateways on the network alongside your twin ID. Ex. ***fw100myfunk*.gent02.dev.grid.tf**
+
+- Enter administrator information including **Username**, **Email** and **Password**. This admin user will have full permission on the deployed instance.
+
+- Select a capacity package:
+ - **Small**: {cpu: 1, memory: 2, diskSize: 50 }
+ - **Medium**: {cpu: 2, memory: 4, diskSize: 100 }
+ - **Large**: {cpu: 4, memory: 16, diskSize: 250 }
+ - Or choose a **Custom** plan
+- Choose the network
+ - `Public IPv4` flag gives the virtual machine a Public IPv4
+
+- `Dedicated` flag to retrieve only dedeicated nodes
+- `Certified` flag to retrieve only certified nodes
+- Choose the location of the node
+ - `Region`
+ - `Country`
+ - `Farm Name`
+
+- Choose the node to deploy on
+> Or you can select a specific node with manual selection.
+- `Custom Domain` flag lets the user to use a custom domain
+- Choose a gateway node to deploy your Funkwhale instance on.
+
+
+After that is done you can see a list of all of your deployed instances
+
+![ ](./img/funkwhale2.png)
+
+Click on ***Visit*** to go to the homepage of your Funkwhale instance!
+
+![ ](./img/funkwhale3.png)
\ No newline at end of file
diff --git a/collections/documentation/dashboard/solutions/img/algorand_defaulttest.png b/collections/documentation/dashboard/solutions/img/algorand_defaulttest.png
new file mode 100644
index 0000000..8fbad1c
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/algorand_defaulttest.png differ
diff --git a/collections/documentation/dashboard/solutions/img/algorand_indexernode.png b/collections/documentation/dashboard/solutions/img/algorand_indexernode.png
new file mode 100644
index 0000000..54540ab
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/algorand_indexernode.png differ
diff --git a/collections/documentation/dashboard/solutions/img/algorand_partdep.png b/collections/documentation/dashboard/solutions/img/algorand_partdep.png
new file mode 100644
index 0000000..bde2f8c
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/algorand_partdep.png differ
diff --git a/collections/documentation/dashboard/solutions/img/algorand_partexp.png b/collections/documentation/dashboard/solutions/img/algorand_partexp.png
new file mode 100644
index 0000000..1c452fb
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/algorand_partexp.png differ
diff --git a/collections/documentation/dashboard/solutions/img/algorand_partonl.png b/collections/documentation/dashboard/solutions/img/algorand_partonl.png
new file mode 100644
index 0000000..984ebd5
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/algorand_partonl.png differ
diff --git a/collections/documentation/dashboard/solutions/img/algorand_partstatus.png b/collections/documentation/dashboard/solutions/img/algorand_partstatus.png
new file mode 100644
index 0000000..f8a9635
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/algorand_partstatus.png differ
diff --git a/collections/documentation/dashboard/solutions/img/algorand_relaydep.png b/collections/documentation/dashboard/solutions/img/algorand_relaydep.png
new file mode 100644
index 0000000..eb04aca
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/algorand_relaydep.png differ
diff --git a/collections/documentation/dashboard/solutions/img/algorand_relaytest.png b/collections/documentation/dashboard/solutions/img/algorand_relaytest.png
new file mode 100644
index 0000000..870b06e
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/algorand_relaytest.png differ
diff --git a/collections/documentation/dashboard/solutions/img/caprover_1.png b/collections/documentation/dashboard/solutions/img/caprover_1.png
new file mode 100644
index 0000000..b6b3f39
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/caprover_1.png differ
diff --git a/collections/documentation/dashboard/solutions/img/caprover_add_node2.png b/collections/documentation/dashboard/solutions/img/caprover_add_node2.png
new file mode 100644
index 0000000..e3ae2a0
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/caprover_add_node2.png differ
diff --git a/collections/documentation/dashboard/solutions/img/caprover_add_worker.png b/collections/documentation/dashboard/solutions/img/caprover_add_worker.png
new file mode 100644
index 0000000..ee0e83a
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/caprover_add_worker.png differ
diff --git a/collections/documentation/dashboard/solutions/img/caprover_cluster.png b/collections/documentation/dashboard/solutions/img/caprover_cluster.png
new file mode 100644
index 0000000..0ede9c8
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/caprover_cluster.png differ
diff --git a/collections/documentation/dashboard/solutions/img/caprover_deploy_leader.png b/collections/documentation/dashboard/solutions/img/caprover_deploy_leader.png
new file mode 100644
index 0000000..0d757e4
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/caprover_deploy_leader.png differ
diff --git a/collections/documentation/dashboard/solutions/img/caprover_deploy_worker.png b/collections/documentation/dashboard/solutions/img/caprover_deploy_worker.png
new file mode 100644
index 0000000..e4f80ce
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/caprover_deploy_worker.png differ
diff --git a/collections/documentation/dashboard/solutions/img/caprover_deploying.png b/collections/documentation/dashboard/solutions/img/caprover_deploying.png
new file mode 100644
index 0000000..307569c
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/caprover_deploying.png differ
diff --git a/collections/documentation/dashboard/solutions/img/caprover_detail_weblet.png b/collections/documentation/dashboard/solutions/img/caprover_detail_weblet.png
new file mode 100644
index 0000000..4f619ff
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/caprover_detail_weblet.png differ
diff --git a/collections/documentation/dashboard/solutions/img/caprover_docker_default_registry.png b/collections/documentation/dashboard/solutions/img/caprover_docker_default_registry.png
new file mode 100644
index 0000000..0998b1c
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/caprover_docker_default_registry.png differ
diff --git a/collections/documentation/dashboard/solutions/img/caprover_docker_registry.png b/collections/documentation/dashboard/solutions/img/caprover_docker_registry.png
new file mode 100644
index 0000000..06abbeb
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/caprover_docker_registry.png differ
diff --git a/collections/documentation/dashboard/solutions/img/caprover_https_activated.png b/collections/documentation/dashboard/solutions/img/caprover_https_activated.png
new file mode 100644
index 0000000..2cbadd6
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/caprover_https_activated.png differ
diff --git a/collections/documentation/dashboard/solutions/img/caprover_login.png b/collections/documentation/dashboard/solutions/img/caprover_login.png
new file mode 100644
index 0000000..3d27d7e
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/caprover_login.png differ
diff --git a/collections/documentation/dashboard/solutions/img/caprover_monitoring_2_.png b/collections/documentation/dashboard/solutions/img/caprover_monitoring_2_.png
new file mode 100644
index 0000000..c92bf01
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/caprover_monitoring_2_.png differ
diff --git a/collections/documentation/dashboard/solutions/img/caprover_monitoring_start_.png b/collections/documentation/dashboard/solutions/img/caprover_monitoring_start_.png
new file mode 100644
index 0000000..2a5fa78
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/caprover_monitoring_start_.png differ
diff --git a/collections/documentation/dashboard/solutions/img/caprover_node_added.png b/collections/documentation/dashboard/solutions/img/caprover_node_added.png
new file mode 100644
index 0000000..93e9a35
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/caprover_node_added.png differ
diff --git a/collections/documentation/dashboard/solutions/img/caprover_select_node.png b/collections/documentation/dashboard/solutions/img/caprover_select_node.png
new file mode 100644
index 0000000..ce155b1
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/caprover_select_node.png differ
diff --git a/collections/documentation/dashboard/solutions/img/captain_login+weblet_caprover_.png b/collections/documentation/dashboard/solutions/img/captain_login+weblet_caprover_.png
new file mode 100644
index 0000000..c4f14d9
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/captain_login+weblet_caprover_.png differ
diff --git a/collections/documentation/dashboard/solutions/img/casper2.png b/collections/documentation/dashboard/solutions/img/casper2.png
new file mode 100644
index 0000000..4948b79
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/casper2.png differ
diff --git a/collections/documentation/dashboard/solutions/img/casper3.png b/collections/documentation/dashboard/solutions/img/casper3.png
new file mode 100644
index 0000000..a9123f0
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/casper3.png differ
diff --git a/collections/documentation/dashboard/solutions/img/casper4.png b/collections/documentation/dashboard/solutions/img/casper4.png
new file mode 100644
index 0000000..17bd144
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/casper4.png differ
diff --git a/collections/documentation/dashboard/solutions/img/casper5.png b/collections/documentation/dashboard/solutions/img/casper5.png
new file mode 100644
index 0000000..5d3f1e7
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/casper5.png differ
diff --git a/collections/documentation/dashboard/solutions/img/cluster_add_nodes.png b/collections/documentation/dashboard/solutions/img/cluster_add_nodes.png
new file mode 100644
index 0000000..c338f54
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/cluster_add_nodes.png differ
diff --git a/collections/documentation/dashboard/solutions/img/contracts_list.png b/collections/documentation/dashboard/solutions/img/contracts_list.png
new file mode 100644
index 0000000..2990bde
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/contracts_list.png differ
diff --git a/collections/documentation/dashboard/solutions/img/deleted_contract_info copy.png b/collections/documentation/dashboard/solutions/img/deleted_contract_info copy.png
new file mode 100644
index 0000000..a3c0357
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/deleted_contract_info copy.png differ
diff --git a/collections/documentation/dashboard/solutions/img/deleted_contract_info.png b/collections/documentation/dashboard/solutions/img/deleted_contract_info.png
new file mode 100644
index 0000000..a3c0357
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/deleted_contract_info.png differ
diff --git a/collections/documentation/dashboard/solutions/img/deplist1.png b/collections/documentation/dashboard/solutions/img/deplist1.png
new file mode 100644
index 0000000..e73d1fd
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/deplist1.png differ
diff --git a/collections/documentation/dashboard/solutions/img/deploy_app_caprover1.png b/collections/documentation/dashboard/solutions/img/deploy_app_caprover1.png
new file mode 100644
index 0000000..7a8fe1e
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/deploy_app_caprover1.png differ
diff --git a/collections/documentation/dashboard/solutions/img/discourse4.png b/collections/documentation/dashboard/solutions/img/discourse4.png
new file mode 100644
index 0000000..eaeb779
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/discourse4.png differ
diff --git a/collections/documentation/dashboard/solutions/img/discourse5.png b/collections/documentation/dashboard/solutions/img/discourse5.png
new file mode 100644
index 0000000..01dca67
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/discourse5.png differ
diff --git a/collections/documentation/dashboard/solutions/img/discourse6.png b/collections/documentation/dashboard/solutions/img/discourse6.png
new file mode 100644
index 0000000..f1566c8
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/discourse6.png differ
diff --git a/collections/documentation/dashboard/solutions/img/domain_name_caprover_config.png b/collections/documentation/dashboard/solutions/img/domain_name_caprover_config.png
new file mode 100644
index 0000000..dca23c3
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/domain_name_caprover_config.png differ
diff --git a/collections/documentation/dashboard/solutions/img/enable_https_caprover.png b/collections/documentation/dashboard/solutions/img/enable_https_caprover.png
new file mode 100644
index 0000000..4149288
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/enable_https_caprover.png differ
diff --git a/collections/documentation/dashboard/solutions/img/fullvm1.png b/collections/documentation/dashboard/solutions/img/fullvm1.png
new file mode 100644
index 0000000..ce820d2
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/fullvm1.png differ
diff --git a/collections/documentation/dashboard/solutions/img/fullvm2.png b/collections/documentation/dashboard/solutions/img/fullvm2.png
new file mode 100644
index 0000000..811bdfc
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/fullvm2.png differ
diff --git a/collections/documentation/dashboard/solutions/img/fullvm4.png b/collections/documentation/dashboard/solutions/img/fullvm4.png
new file mode 100644
index 0000000..5186f74
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/fullvm4.png differ
diff --git a/collections/documentation/dashboard/solutions/img/fullvm5.png b/collections/documentation/dashboard/solutions/img/fullvm5.png
new file mode 100644
index 0000000..376d3d9
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/fullvm5.png differ
diff --git a/collections/documentation/dashboard/solutions/img/fullvm6.png b/collections/documentation/dashboard/solutions/img/fullvm6.png
new file mode 100644
index 0000000..30348fc
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/fullvm6.png differ
diff --git a/collections/documentation/dashboard/solutions/img/fullvm7.jpg b/collections/documentation/dashboard/solutions/img/fullvm7.jpg
new file mode 100644
index 0000000..f5fd320
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/fullvm7.jpg differ
diff --git a/collections/documentation/dashboard/solutions/img/funkwhale2.png b/collections/documentation/dashboard/solutions/img/funkwhale2.png
new file mode 100644
index 0000000..de4fab4
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/funkwhale2.png differ
diff --git a/collections/documentation/dashboard/solutions/img/funkwhale3.png b/collections/documentation/dashboard/solutions/img/funkwhale3.png
new file mode 100644
index 0000000..7a5f00c
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/funkwhale3.png differ
diff --git a/collections/documentation/dashboard/solutions/img/k8s_dl_1.png b/collections/documentation/dashboard/solutions/img/k8s_dl_1.png
new file mode 100644
index 0000000..c8b5670
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/k8s_dl_1.png differ
diff --git a/collections/documentation/dashboard/solutions/img/k8s_dl_2.png b/collections/documentation/dashboard/solutions/img/k8s_dl_2.png
new file mode 100644
index 0000000..a5874f7
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/k8s_dl_2.png differ
diff --git a/collections/documentation/dashboard/solutions/img/k8s_dl_4.png b/collections/documentation/dashboard/solutions/img/k8s_dl_4.png
new file mode 100644
index 0000000..be4361c
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/k8s_dl_4.png differ
diff --git a/collections/documentation/dashboard/solutions/img/mastodon1.jpg b/collections/documentation/dashboard/solutions/img/mastodon1.jpg
new file mode 100644
index 0000000..e4bc3d6
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/mastodon1.jpg differ
diff --git a/collections/documentation/dashboard/solutions/img/mastodon2.jpg b/collections/documentation/dashboard/solutions/img/mastodon2.jpg
new file mode 100644
index 0000000..ae0748b
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/mastodon2.jpg differ
diff --git a/collections/documentation/dashboard/solutions/img/mastodon3.jpg b/collections/documentation/dashboard/solutions/img/mastodon3.jpg
new file mode 100644
index 0000000..591570b
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/mastodon3.jpg differ
diff --git a/collections/documentation/dashboard/solutions/img/mastodon4.jpg b/collections/documentation/dashboard/solutions/img/mastodon4.jpg
new file mode 100644
index 0000000..f9afe85
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/mastodon4.jpg differ
diff --git a/collections/documentation/dashboard/solutions/img/mastodon5.jpg b/collections/documentation/dashboard/solutions/img/mastodon5.jpg
new file mode 100644
index 0000000..eb01556
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/mastodon5.jpg differ
diff --git a/collections/documentation/dashboard/solutions/img/mattermost3.png b/collections/documentation/dashboard/solutions/img/mattermost3.png
new file mode 100644
index 0000000..2ada6dc
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/mattermost3.png differ
diff --git a/collections/documentation/dashboard/solutions/img/mattermost4.png b/collections/documentation/dashboard/solutions/img/mattermost4.png
new file mode 100644
index 0000000..fd6ee19
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/mattermost4.png differ
diff --git a/collections/documentation/dashboard/solutions/img/mattermost5.png b/collections/documentation/dashboard/solutions/img/mattermost5.png
new file mode 100644
index 0000000..a45f632
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/mattermost5.png differ
diff --git a/collections/documentation/dashboard/solutions/img/new_cap2.png b/collections/documentation/dashboard/solutions/img/new_cap2.png
new file mode 100644
index 0000000..689da23
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/new_cap2.png differ
diff --git a/collections/documentation/dashboard/solutions/img/new_cap3.png b/collections/documentation/dashboard/solutions/img/new_cap3.png
new file mode 100644
index 0000000..31c5110
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/new_cap3.png differ
diff --git a/collections/documentation/dashboard/solutions/img/new_funk2.png b/collections/documentation/dashboard/solutions/img/new_funk2.png
new file mode 100644
index 0000000..eb87d9f
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/new_funk2.png differ
diff --git a/collections/documentation/dashboard/solutions/img/new_funk3.png b/collections/documentation/dashboard/solutions/img/new_funk3.png
new file mode 100644
index 0000000..4e97d6d
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/new_funk3.png differ
diff --git a/collections/documentation/dashboard/solutions/img/new_k8s4.png b/collections/documentation/dashboard/solutions/img/new_k8s4.png
new file mode 100644
index 0000000..9dded9c
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/new_k8s4.png differ
diff --git a/collections/documentation/dashboard/solutions/img/new_k8s5.png b/collections/documentation/dashboard/solutions/img/new_k8s5.png
new file mode 100644
index 0000000..e6aa91a
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/new_k8s5.png differ
diff --git a/collections/documentation/dashboard/solutions/img/new_peer2.png b/collections/documentation/dashboard/solutions/img/new_peer2.png
new file mode 100644
index 0000000..604fe10
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/new_peer2.png differ
diff --git a/collections/documentation/dashboard/solutions/img/new_peer3.png b/collections/documentation/dashboard/solutions/img/new_peer3.png
new file mode 100644
index 0000000..43987a4
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/new_peer3.png differ
diff --git a/collections/documentation/dashboard/solutions/img/new_vm2.png b/collections/documentation/dashboard/solutions/img/new_vm2.png
new file mode 100644
index 0000000..c5ebf1d
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/new_vm2.png differ
diff --git a/collections/documentation/dashboard/solutions/img/new_vm3.png b/collections/documentation/dashboard/solutions/img/new_vm3.png
new file mode 100644
index 0000000..1981cb6
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/new_vm3.png differ
diff --git a/collections/documentation/dashboard/solutions/img/new_vm5.png b/collections/documentation/dashboard/solutions/img/new_vm5.png
new file mode 100644
index 0000000..723a44e
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/new_vm5.png differ
diff --git a/collections/documentation/dashboard/solutions/img/new_vm6.png b/collections/documentation/dashboard/solutions/img/new_vm6.png
new file mode 100644
index 0000000..d2af47d
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/new_vm6.png differ
diff --git a/collections/documentation/dashboard/solutions/img/new_vm7.png b/collections/documentation/dashboard/solutions/img/new_vm7.png
new file mode 100644
index 0000000..da8c92b
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/new_vm7.png differ
diff --git a/collections/documentation/dashboard/solutions/img/new_vm8.png b/collections/documentation/dashboard/solutions/img/new_vm8.png
new file mode 100644
index 0000000..ee291c8
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/new_vm8.png differ
diff --git a/collections/documentation/dashboard/solutions/img/nixos-micro1.png b/collections/documentation/dashboard/solutions/img/nixos-micro1.png
new file mode 100644
index 0000000..0114907
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/nixos-micro1.png differ
diff --git a/collections/documentation/dashboard/solutions/img/nixos-micro2.png b/collections/documentation/dashboard/solutions/img/nixos-micro2.png
new file mode 100644
index 0000000..d094ef3
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/nixos-micro2.png differ
diff --git a/collections/documentation/dashboard/solutions/img/nixos-micro3.png b/collections/documentation/dashboard/solutions/img/nixos-micro3.png
new file mode 100644
index 0000000..85247c3
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/nixos-micro3.png differ
diff --git a/collections/documentation/dashboard/solutions/img/nodeP_2.png b/collections/documentation/dashboard/solutions/img/nodeP_2.png
new file mode 100644
index 0000000..f1f7ac2
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/nodeP_2.png differ
diff --git a/collections/documentation/dashboard/solutions/img/nodePilot_2.png b/collections/documentation/dashboard/solutions/img/nodePilot_2.png
new file mode 100644
index 0000000..083551b
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/nodePilot_2.png differ
diff --git a/collections/documentation/dashboard/solutions/img/nodePilot_3.png b/collections/documentation/dashboard/solutions/img/nodePilot_3.png
new file mode 100644
index 0000000..8b4549b
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/nodePilot_3.png differ
diff --git a/collections/documentation/dashboard/solutions/img/node_selection.png b/collections/documentation/dashboard/solutions/img/node_selection.png
new file mode 100644
index 0000000..50a30dc
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/node_selection.png differ
diff --git a/collections/documentation/dashboard/solutions/img/nxios-micro1.png b/collections/documentation/dashboard/solutions/img/nxios-micro1.png
new file mode 100644
index 0000000..c2fc9d7
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/nxios-micro1.png differ
diff --git a/collections/documentation/dashboard/solutions/img/owncloud1.png b/collections/documentation/dashboard/solutions/img/owncloud1.png
new file mode 100644
index 0000000..2a5567e
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/owncloud1.png differ
diff --git a/collections/documentation/dashboard/solutions/img/owncloud2.png b/collections/documentation/dashboard/solutions/img/owncloud2.png
new file mode 100644
index 0000000..c34fab7
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/owncloud2.png differ
diff --git a/collections/documentation/dashboard/solutions/img/owncloud3.png b/collections/documentation/dashboard/solutions/img/owncloud3.png
new file mode 100644
index 0000000..62995d3
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/owncloud3.png differ
diff --git a/collections/documentation/dashboard/solutions/img/owncloud4.png b/collections/documentation/dashboard/solutions/img/owncloud4.png
new file mode 100644
index 0000000..98d16ae
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/owncloud4.png differ
diff --git a/collections/documentation/dashboard/solutions/img/owncloud5.png b/collections/documentation/dashboard/solutions/img/owncloud5.png
new file mode 100644
index 0000000..d441b86
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/owncloud5.png differ
diff --git a/collections/documentation/dashboard/solutions/img/owncloud6.png b/collections/documentation/dashboard/solutions/img/owncloud6.png
new file mode 100644
index 0000000..a994782
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/owncloud6.png differ
diff --git a/collections/documentation/dashboard/solutions/img/presearch0.png b/collections/documentation/dashboard/solutions/img/presearch0.png
new file mode 100644
index 0000000..29dedbe
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/presearch0.png differ
diff --git a/collections/documentation/dashboard/solutions/img/presearch4.png b/collections/documentation/dashboard/solutions/img/presearch4.png
new file mode 100644
index 0000000..2c3c8e9
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/presearch4.png differ
diff --git a/collections/documentation/dashboard/solutions/img/presearch5.png b/collections/documentation/dashboard/solutions/img/presearch5.png
new file mode 100644
index 0000000..9c7ec67
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/presearch5.png differ
diff --git a/collections/documentation/dashboard/solutions/img/presearch6.png b/collections/documentation/dashboard/solutions/img/presearch6.png
new file mode 100644
index 0000000..9279f30
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/presearch6.png differ
diff --git a/collections/documentation/dashboard/solutions/img/qvm1.png b/collections/documentation/dashboard/solutions/img/qvm1.png
new file mode 100644
index 0000000..93277ab
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/qvm1.png differ
diff --git a/collections/documentation/dashboard/solutions/img/qvm2.png b/collections/documentation/dashboard/solutions/img/qvm2.png
new file mode 100644
index 0000000..0777158
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/qvm2.png differ
diff --git a/collections/documentation/dashboard/solutions/img/qvm_nodes.png b/collections/documentation/dashboard/solutions/img/qvm_nodes.png
new file mode 100644
index 0000000..3670785
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/qvm_nodes.png differ
diff --git a/collections/documentation/dashboard/solutions/img/qvm_qsfs_config.png b/collections/documentation/dashboard/solutions/img/qvm_qsfs_config.png
new file mode 100644
index 0000000..6e337e4
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/qvm_qsfs_config.png differ
diff --git a/collections/documentation/dashboard/solutions/img/solution_caprover_list.png b/collections/documentation/dashboard/solutions/img/solution_caprover_list.png
new file mode 100644
index 0000000..3f672e2
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/solution_caprover_list.png differ
diff --git a/collections/documentation/dashboard/solutions/img/solutions_algorand.png b/collections/documentation/dashboard/solutions/img/solutions_algorand.png
new file mode 100644
index 0000000..ca7e860
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/solutions_algorand.png differ
diff --git a/collections/documentation/dashboard/solutions/img/solutions_caprover.png b/collections/documentation/dashboard/solutions/img/solutions_caprover.png
new file mode 100644
index 0000000..bf2f072
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/solutions_caprover.png differ
diff --git a/collections/documentation/dashboard/solutions/img/solutions_caprover_leader.png b/collections/documentation/dashboard/solutions/img/solutions_caprover_leader.png
new file mode 100644
index 0000000..218931d
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/solutions_caprover_leader.png differ
diff --git a/collections/documentation/dashboard/solutions/img/solutions_caprover_workers.png b/collections/documentation/dashboard/solutions/img/solutions_caprover_workers.png
new file mode 100644
index 0000000..0b12a69
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/solutions_caprover_workers.png differ
diff --git a/collections/documentation/dashboard/solutions/img/solutions_casperlabs.png b/collections/documentation/dashboard/solutions/img/solutions_casperlabs.png
new file mode 100644
index 0000000..f9f3271
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/solutions_casperlabs.png differ
diff --git a/collections/documentation/dashboard/solutions/img/solutions_discourse.png b/collections/documentation/dashboard/solutions/img/solutions_discourse.png
new file mode 100644
index 0000000..d7e0c66
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/solutions_discourse.png differ
diff --git a/collections/documentation/dashboard/solutions/img/solutions_fullvm.png b/collections/documentation/dashboard/solutions/img/solutions_fullvm.png
new file mode 100644
index 0000000..32d786d
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/solutions_fullvm.png differ
diff --git a/collections/documentation/dashboard/solutions/img/solutions_funkwhale.png b/collections/documentation/dashboard/solutions/img/solutions_funkwhale.png
new file mode 100644
index 0000000..6802db6
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/solutions_funkwhale.png differ
diff --git a/collections/documentation/dashboard/solutions/img/solutions_k8s.png b/collections/documentation/dashboard/solutions/img/solutions_k8s.png
new file mode 100644
index 0000000..935de58
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/solutions_k8s.png differ
diff --git a/collections/documentation/dashboard/solutions/img/solutions_k8s_master.png b/collections/documentation/dashboard/solutions/img/solutions_k8s_master.png
new file mode 100644
index 0000000..948a43b
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/solutions_k8s_master.png differ
diff --git a/collections/documentation/dashboard/solutions/img/solutions_k8s_workers.png b/collections/documentation/dashboard/solutions/img/solutions_k8s_workers.png
new file mode 100644
index 0000000..5e6df13
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/solutions_k8s_workers.png differ
diff --git a/collections/documentation/dashboard/solutions/img/solutions_mattermost.png b/collections/documentation/dashboard/solutions/img/solutions_mattermost.png
new file mode 100644
index 0000000..e47bb61
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/solutions_mattermost.png differ
diff --git a/collections/documentation/dashboard/solutions/img/solutions_microvm.png b/collections/documentation/dashboard/solutions/img/solutions_microvm.png
new file mode 100644
index 0000000..abad7ff
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/solutions_microvm.png differ
diff --git a/collections/documentation/dashboard/solutions/img/solutions_nextcloud.png b/collections/documentation/dashboard/solutions/img/solutions_nextcloud.png
new file mode 100644
index 0000000..a588f97
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/solutions_nextcloud.png differ
diff --git a/collections/documentation/dashboard/solutions/img/solutions_nodepilot.png b/collections/documentation/dashboard/solutions/img/solutions_nodepilot.png
new file mode 100644
index 0000000..65de04c
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/solutions_nodepilot.png differ
diff --git a/collections/documentation/dashboard/solutions/img/solutions_owncloud.png b/collections/documentation/dashboard/solutions/img/solutions_owncloud.png
new file mode 100644
index 0000000..98f7302
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/solutions_owncloud.png differ
diff --git a/collections/documentation/dashboard/solutions/img/solutions_owncloud_visit.png b/collections/documentation/dashboard/solutions/img/solutions_owncloud_visit.png
new file mode 100644
index 0000000..45c8857
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/solutions_owncloud_visit.png differ
diff --git a/collections/documentation/dashboard/solutions/img/solutions_peertube.png b/collections/documentation/dashboard/solutions/img/solutions_peertube.png
new file mode 100644
index 0000000..341653d
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/solutions_peertube.png differ
diff --git a/collections/documentation/dashboard/solutions/img/solutions_presearch.png b/collections/documentation/dashboard/solutions/img/solutions_presearch.png
new file mode 100644
index 0000000..258fca0
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/solutions_presearch.png differ
diff --git a/collections/documentation/dashboard/solutions/img/solutions_subsquid.png b/collections/documentation/dashboard/solutions/img/solutions_subsquid.png
new file mode 100644
index 0000000..2b2f132
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/solutions_subsquid.png differ
diff --git a/collections/documentation/dashboard/solutions/img/solutions_taiga.png b/collections/documentation/dashboard/solutions/img/solutions_taiga.png
new file mode 100644
index 0000000..b4f89e1
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/solutions_taiga.png differ
diff --git a/collections/documentation/dashboard/solutions/img/solutions_umbrel.png b/collections/documentation/dashboard/solutions/img/solutions_umbrel.png
new file mode 100644
index 0000000..311d83c
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/solutions_umbrel.png differ
diff --git a/collections/documentation/dashboard/solutions/img/solutions_wordpress.png b/collections/documentation/dashboard/solutions/img/solutions_wordpress.png
new file mode 100644
index 0000000..58e66b2
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/solutions_wordpress.png differ
diff --git a/collections/documentation/dashboard/solutions/img/subsquid_graphql.png b/collections/documentation/dashboard/solutions/img/subsquid_graphql.png
new file mode 100644
index 0000000..82cf6d0
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/subsquid_graphql.png differ
diff --git a/collections/documentation/dashboard/solutions/img/subsquid_list.jpeg b/collections/documentation/dashboard/solutions/img/subsquid_list.jpeg
new file mode 100644
index 0000000..596497f
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/subsquid_list.jpeg differ
diff --git a/collections/documentation/dashboard/solutions/img/subsquid_list.png b/collections/documentation/dashboard/solutions/img/subsquid_list.png
new file mode 100644
index 0000000..c0a78aa
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/subsquid_list.png differ
diff --git a/collections/documentation/dashboard/solutions/img/taiga2.png b/collections/documentation/dashboard/solutions/img/taiga2.png
new file mode 100644
index 0000000..cdcfcd4
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/taiga2.png differ
diff --git a/collections/documentation/dashboard/solutions/img/taiga3.png b/collections/documentation/dashboard/solutions/img/taiga3.png
new file mode 100644
index 0000000..35ddd01
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/taiga3.png differ
diff --git a/collections/documentation/dashboard/solutions/img/taiga4.png b/collections/documentation/dashboard/solutions/img/taiga4.png
new file mode 100644
index 0000000..8b9c844
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/taiga4.png differ
diff --git a/collections/documentation/dashboard/solutions/img/taiga5.png b/collections/documentation/dashboard/solutions/img/taiga5.png
new file mode 100644
index 0000000..f0bcc2b
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/taiga5.png differ
diff --git a/collections/documentation/dashboard/solutions/img/taiga6.png b/collections/documentation/dashboard/solutions/img/taiga6.png
new file mode 100644
index 0000000..606939b
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/taiga6.png differ
diff --git a/collections/documentation/dashboard/solutions/img/umbrel2.png b/collections/documentation/dashboard/solutions/img/umbrel2.png
new file mode 100644
index 0000000..859e141
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/umbrel2.png differ
diff --git a/collections/documentation/dashboard/solutions/img/umbrel3.png b/collections/documentation/dashboard/solutions/img/umbrel3.png
new file mode 100644
index 0000000..c8dacd5
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/umbrel3.png differ
diff --git a/collections/documentation/dashboard/solutions/img/umbrel4.png b/collections/documentation/dashboard/solutions/img/umbrel4.png
new file mode 100644
index 0000000..1d9bdf1
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/umbrel4.png differ
diff --git a/collections/documentation/dashboard/solutions/img/umbrel5.png b/collections/documentation/dashboard/solutions/img/umbrel5.png
new file mode 100644
index 0000000..4d7379d
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/umbrel5.png differ
diff --git a/collections/documentation/dashboard/solutions/img/vm_json.png b/collections/documentation/dashboard/solutions/img/vm_json.png
new file mode 100644
index 0000000..1ab3737
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/vm_json.png differ
diff --git a/collections/documentation/dashboard/solutions/img/vm_list.png b/collections/documentation/dashboard/solutions/img/vm_list.png
new file mode 100644
index 0000000..2a504b1
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/vm_list.png differ
diff --git a/collections/documentation/dashboard/solutions/img/weblet_peertube_instance.png b/collections/documentation/dashboard/solutions/img/weblet_peertube_instance.png
new file mode 100644
index 0000000..4474a75
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/weblet_peertube_instance.png differ
diff --git a/collections/documentation/dashboard/solutions/img/weblet_peertube_listing.png b/collections/documentation/dashboard/solutions/img/weblet_peertube_listing.png
new file mode 100644
index 0000000..93de06f
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/weblet_peertube_listing.png differ
diff --git a/collections/documentation/dashboard/solutions/img/weblet_vm4.png b/collections/documentation/dashboard/solutions/img/weblet_vm4.png
new file mode 100644
index 0000000..f088303
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/weblet_vm4.png differ
diff --git a/collections/documentation/dashboard/solutions/img/weblet_vm5.png b/collections/documentation/dashboard/solutions/img/weblet_vm5.png
new file mode 100644
index 0000000..0f778f5
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/weblet_vm5.png differ
diff --git a/collections/documentation/dashboard/solutions/img/weblet_vm_overview.png b/collections/documentation/dashboard/solutions/img/weblet_vm_overview.png
new file mode 100644
index 0000000..3262552
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/weblet_vm_overview.png differ
diff --git a/collections/documentation/dashboard/solutions/img/weblet_vm_presearch_result.jpg b/collections/documentation/dashboard/solutions/img/weblet_vm_presearch_result.jpg
new file mode 100644
index 0000000..ae77142
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/weblet_vm_presearch_result.jpg differ
diff --git a/collections/documentation/dashboard/solutions/img/wp10.png b/collections/documentation/dashboard/solutions/img/wp10.png
new file mode 100644
index 0000000..c4f9998
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/wp10.png differ
diff --git a/collections/documentation/dashboard/solutions/img/wp11.png b/collections/documentation/dashboard/solutions/img/wp11.png
new file mode 100644
index 0000000..35e9f58
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/wp11.png differ
diff --git a/collections/documentation/dashboard/solutions/img/wp2.png b/collections/documentation/dashboard/solutions/img/wp2.png
new file mode 100644
index 0000000..8e3aed3
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/wp2.png differ
diff --git a/collections/documentation/dashboard/solutions/img/wp3.png b/collections/documentation/dashboard/solutions/img/wp3.png
new file mode 100644
index 0000000..e485277
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/wp3.png differ
diff --git a/collections/documentation/dashboard/solutions/img/wp4.png b/collections/documentation/dashboard/solutions/img/wp4.png
new file mode 100644
index 0000000..8d534f1
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/wp4.png differ
diff --git a/collections/documentation/dashboard/solutions/img/wp5.png b/collections/documentation/dashboard/solutions/img/wp5.png
new file mode 100644
index 0000000..141497d
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/wp5.png differ
diff --git a/collections/documentation/dashboard/solutions/img/wp6.png b/collections/documentation/dashboard/solutions/img/wp6.png
new file mode 100644
index 0000000..0f72d48
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/wp6.png differ
diff --git a/collections/documentation/dashboard/solutions/img/wp7.png b/collections/documentation/dashboard/solutions/img/wp7.png
new file mode 100644
index 0000000..01af952
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/wp7.png differ
diff --git a/collections/documentation/dashboard/solutions/img/wp8.png b/collections/documentation/dashboard/solutions/img/wp8.png
new file mode 100644
index 0000000..f243cd2
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/wp8.png differ
diff --git a/collections/documentation/dashboard/solutions/img/wp9.png b/collections/documentation/dashboard/solutions/img/wp9.png
new file mode 100644
index 0000000..34173dd
Binary files /dev/null and b/collections/documentation/dashboard/solutions/img/wp9.png differ
diff --git a/collections/documentation/dashboard/solutions/k8s.md b/collections/documentation/dashboard/solutions/k8s.md
new file mode 100644
index 0000000..11197ec
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/k8s.md
@@ -0,0 +1,98 @@
+ Kubernetes
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Prerequisites](#prerequisites)
+- [Configs tab](#configs-tab)
+- [Master and Workers tabs](#master-and-workers-tabs)
+- [Kubeconfig](#kubeconfig)
+- [Manage Workers](#manage-workers)
+
+***
+
+## Introduction
+
+Kubernetes is the standard container orchestration tool.
+
+On the TF grid, Kubernetes clusters can be deployed out of the box. We have implemented [K3S](https://k3s.io/), a full-blown Kubernetes offering that uses only half of the memory footprint. It is packaged as a single binary and made more lightweight to run workloads in resource-constrained locations (fits e.g. IoT, edge, ARM workloads).
+
+## Prerequisites
+
+- Make sure you have a [wallet](../wallet_connector.md)
+- From the sidebar click on **Applications**
+- Click on **Kubernetes**
+
+## Configs tab
+
+![ ](./img/solutions_k8s.png)
+
+- `Name`: Your Kubernetes Cluster name.
+- `Cluster Token`: It's used for authentication between your worker nodes and master node. You could use the auto-generated one or type your own.
+
+
+## Master and Workers tabs
+
+![ ](./img/solutions_k8s_master.png)
+![ ](./img/solutions_k8s_workers.png)
+
+> Currently, we only support "single-master-multi-worker" k8s clusters. So you could always add more than one worker node by clicking on the **+** in the ***Worker*** tab.
+
+
+## Kubeconfig
+Once the cluster is ready, you can SSH into the cluster using `ssh root@IP`
+> IP can be the public IP or the planetary network IP
+
+Onced connected via SSH, you can execute commands on the cluster like `kubectl get nodes`, and to get the kubeconfig, you can find it in `/root/.kube/config`
+
+> if it doesn't exist in `/root/.kube/config` it can be in `/etc/rancher/k3s/k3s.yaml`
+
+example:
+
+```
+root@WR768dbf76:~# cat /root/.kube/config
+apiVersion: v1
+clusters:
+- cluster:
+ certificate-authority-data: LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSUJkakNDQVIyZ0F3SUJBZ0lCQURBS0JnZ3Foa2pPUFFRREFqQWpNU0V3SHdZRFZRUUREQmhyTTNNdGMyVnkKZG1WeUxXTmhRREUyTkRBeU5qWTBNVE13SGhjTk1qRXhNakl6TVRNek16TXpXaGNOTXpFeE1qSXhNVE16TXpNegpXakFqTVNFd0h3WURWUVFEREJock0zTXRjMlZ5ZG1WeUxXTmhRREUyTkRBeU5qWTBNVE13V1RBVEJnY3Foa2pPClBRSUJCZ2dxaGtqT1BRTUJCd05DQUFUcGNtZE1KaWg1eGFTa1JlelNKVU5mUkQ5NWV6cE12amhVeUc2bWU4bTkKY0lQWENoNUZ2ZU81Znk1d1VTSTlYOFlGV2JkOGtRcG9vaVdVbStwYjFvU3hvMEl3UURBT0JnTlZIUThCQWY4RQpCQU1DQXFRd0R3WURWUjBUQVFIL0JBVXdBd0VCL3pBZEJnTlZIUTRFRmdRVUtkL3VUU3FtWk12bHhtcWNYU3lxCmVhWERIbXd3Q2dZSUtvWkl6ajBFQXdJRFJ3QXdSQUlnSUQ0cGNQWDl2R0F6SC9lTkhCNndVdmNZRi9HbXFuQVIKR2dqT1RSdWVia1lDSUdRUmUwTGJzQXdwMWNicHlYRWljV3V0aG1RQ1dwRXY1NThWZ3BoMFpETFAKLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=
+ server: https://127.0.0.1:6443
+ name: default
+contexts:
+- context:
+ cluster: default
+ user: default
+ name: default
+current-context: default
+kind: Config
+preferences: {}
+users:
+- name: default
+ user:
+ client-certificate-data: LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSUJrVENDQVRlZ0F3SUJBZ0lJWnptV1A4ellKaGd3Q2dZSUtvWkl6ajBFQXdJd0l6RWhNQjhHQTFVRUF3d1kKYXpOekxXTnNhV1Z1ZEMxallVQXhOalF3TWpZMk5ERXpNQjRYRFRJeE1USXlNekV6TXpNek0xb1hEVEl5TVRJeQpNekV6TXpNek0xb3dNREVYTUJVR0ExVUVDaE1PYzNsemRHVnRPbTFoYzNSbGNuTXhGVEFUQmdOVkJBTVRESE41CmMzUmxiVHBoWkcxcGJqQlpNQk1HQnlxR1NNNDlBZ0VHQ0NxR1NNNDlBd0VIQTBJQUJINTZaZGM5aTJ0azAyNGQKcXBDQ2NRMndMMjc1QWtPZUFxalIzQjFTTGFQeG1oOG9IcXd4SzY2RTc1ZWQya2VySFIySnBZbWwwNE5sa0grLwpSd2kvMDNDalNEQkdNQTRHQTFVZER3RUIvd1FFQXdJRm9EQVRCZ05WSFNVRUREQUtCZ2dyQmdFRkJRY0RBakFmCkJnTlZIU01FR0RBV2dCVGhPakJSaExjeE53UDkzd0xtUzBYRUFUNjlSekFLQmdncWhrak9QUVFEQWdOSUFEQkYKQWlBcjdDcDR2dks4Y2s0Q0lROEM5em5zVkFUZVhDaHZsUmdvanZuVXU4REZld0loQUlwRVYyMWJZVXBpUEkzVQowa3QvQmJqRUtjV1poVXNHQ0g0YzVNWTFFS0JhCi0tLS0tRU5EIENFUlRJRklDQVRFLS0tLS0KLS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSUJkekNDQVIyZ0F3SUJBZ0lCQURBS0JnZ3Foa2pPUFFRREFqQWpNU0V3SHdZRFZRUUREQmhyTTNNdFkyeHAKWlc1MExXTmhRREUyTkRBeU5qWTBNVE13SGhjTk1qRXhNakl6TVRNek16TXpXaGNOTXpFeE1qSXhNVE16TXpNegpXakFqTVNFd0h3WURWUVFEREJock0zTXRZMnhwWlc1MExXTmhRREUyTkRBeU5qWTBNVE13V1RBVEJnY3Foa2pPClBRSUJCZ2dxaGtqT1BRTUJCd05DQUFUY3NlakN3TDQ5VkZvQnJhWVRyR3ByR2lMajNKeEw4ZVcwYnpTVDBWRGUKeFlrb3hDbDlnR0N6R2p1Q2Q0ZmZmRXV0QWdFMjU5MDFBWGJCU2VnOHdlSkJvMEl3UURBT0JnTlZIUThCQWY4RQpCQU1DQXFRd0R3WURWUjBUQVFIL0JBVXdBd0VCL3pBZEJnTlZIUTRFRmdRVTRUb3dVWVMzTVRjRC9kOEM1a3RGCnhBRSt2VWN3Q2dZSUtvWkl6ajBFQXdJRFNBQXdSUUloQU5CYWRhcFFZbnlYOEJDUllNODZtYWtMNkFDM0hSenMKL2l3Ukp6TnV6YytaQWlCZm14YytDTVZHQnBrblAzR2dWSWlFMFVQWkUrOFRnRUdkTTgrdCt4V2Ywdz09Ci0tLS0tRU5EIENFUlRJRklDQVRFLS0tLS0K
+ client-key-data: LS0tLS1CRUdJTiBFQyBQUklWQVRFIEtFWS0tLS0tCk1IY0NBUUVFSURXQURoZUl0RVdHWlFCc0tCSUpZTTZPeDB5TmRHQ1JjTDBTMUtvYjRTZ25vQW9HQ0NxR1NNNDkKQXdFSG9VUURRZ0FFZm5wbDF6MkxhMlRUYmgycWtJSnhEYkF2YnZrQ1E1NENxTkhjSFZJdG8vR2FIeWdlckRFcgpyb1R2bDUzYVI2c2RIWW1saWFYVGcyV1FmNzlIQ0wvVGNBPT0KLS0tLS1FTkQgRUMgUFJJVkFURSBLRVktLS0tLQo=
+root@WR768dbf76:~#
+
+```
+
+If you want to use kubectl through another machine, you need to change the line `server: https://127.0.0.1:6443` to be `server: https://PLANETARYIP_OR_PUBLICIP/6443`
+replace PLANETARYIP_OR_PUBLICIP with the IP you want to reach th cluster through.
+
+
+## Manage Workers
+Add or Remove workers in any **Kubernetes cluster**.
+
+
+- Kubernetes DeployedList Weblet
+![ ](./img/k8s_dl_1.png)
+
+- Manager kubernetes workers
+![ ](./img/k8s_dl_2.png)
+
+- Add a new worker
+![ ](./img/new_k8s4.png)
+
+- Successfully added new worker
+![ ](./img/k8s_dl_4.png)
+
+- Delete a worker
+![ ](./img/new_k8s5.png)
\ No newline at end of file
diff --git a/collections/documentation/dashboard/solutions/mattermost.md b/collections/documentation/dashboard/solutions/mattermost.md
new file mode 100644
index 0000000..0e9528c
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/mattermost.md
@@ -0,0 +1,55 @@
+ Mattermost
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Prerequisites](#prerequisites)
+- [Deployment](#deployment)
+
+***
+
+## Introduction
+
+[Mattermost](https://mattermost.com/) A single point of collaboration. Designed specifically for digital operations.
+
+## Prerequisites
+
+- Make sure you have a [wallet](../wallet_connector.md)
+- From the sidebar click on **Applications**
+- Click on **Mattermost**
+
+## Deployment
+
+![ ](./img/solutions_mattermost.png)
+
+- Enter an Application Name. It's used in generating a unique subdomain on one of the gateways on the network alongside your twin ID. Ex. ***matter*.gent02.dev.grid.tf**
+
+- Select a capacity package:
+ - **Small**: {cpu: 1, memory: 2, diskSize: 15 }
+ - **Medium**: {cpu: 2, memory: 4, diskSize: 50 }
+ - **Large**: {cpu: 4, memory: 16, diskSize: 100 }
+ - Or choose a **Custom** plan
+- `Dedicated` flag to retrieve only dedeicated nodes
+- `Certified` flag to retrieve only certified nodes
+- Choose the location of the node
+ - `Region`
+ - `Country`
+ - `Farm Name`
+
+- Choose the node to deploy on
+> Or you can select a specific node with manual selection.
+- `Custom Domain` flag lets the user to use a custom domain
+- Choose a gateway node to deploy your Mattermost instance on.
+
+
+- There's also an optional **SMTP Server** tab if you'd like to have your Mattermost instance configured with an SMTP server.
+
+ ![ ](./img/mattermost3.png)
+
+After that is done you can see a list of all of your deployed instances
+
+![ ](./img/mattermost4.png)
+
+Click on ***Visit*** to go to the homepage of your Mattermost instance! You need to login using TFConnect so make sure you download the *TFConnect* app from your App Store.
+
+![ ](./img/mattermost5.png)
diff --git a/collections/documentation/dashboard/solutions/nextcloud.md b/collections/documentation/dashboard/solutions/nextcloud.md
new file mode 100644
index 0000000..d25dd79
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/nextcloud.md
@@ -0,0 +1,207 @@
+ Nextcloud
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Prerequisites](#prerequisites)
+- [Domain Names and Public IPs](#domain-names-and-public-ips)
+- [Deploy Nextcloud](#deploy-nextcloud)
+- [Nextcloud Setup](#nextcloud-setup)
+- [DNS Details](#dns-details)
+ - [DNS Record with Public IPv4](#dns-record-with-public-ipv4)
+ - [DNS Record with Gateway](#dns-record-with-gateway)
+ - [DNS Propagation](#dns-propagation)
+- [Talk](#talk)
+ - [Install Talk](#install-talk)
+ - [TURN](#turn)
+ - [Use Talk](#use-talk)
+- [Backups and Updates](#backups-and-updates)
+ - [Create a Backup](#create-a-backup)
+ - [Automatic Backups and Updates](#automatic-backups-and-updates)
+- [Troubleshooting](#troubleshooting)
+ - [Retrieve the Nextcloud AIO Password](#retrieve-the-nextcloud-aio-password)
+ - [Access the Nextcloud Interface Page](#access-the-nextcloud-interface-page)
+ - [Check the DNS Propagation](#check-the-dns-propagation)
+- [Questions and Feedback](#questions-and-feedback)
+
+***
+
+# Introduction
+
+[Nextcloud](https://nextcloud.com/) is a suite of client-server software for creating and using file hosting services.
+
+Nextcloud provides functionality similar to Dropbox, Office 365 or Google Drive when used with integrated office suites like Collabora Online or OnlyOffice.
+
+
+
+# Prerequisites
+
+- Make sure you have a [wallet](../wallet_connector.md)
+- From the sidebar click on **Applications**
+- Click on **Nextcloud**
+
+
+
+# Domain Names and Public IPs
+
+A domain name is required to use Nextcloud. You can either use your own, which we'll call a *custom domain*, or you can get a free subdomain from a gateway node. This won't impact the function of your deployment, it's just a matter of preference. If you want to use your own domain, follow the steps for custom domain wherever you see them below.
+
+Another choice to make before launching your Nextcloud instance is whether you want to reserve a public IPv4 for the deployment. Renting a public IP is an extra cost and is only required for the dedicated Nextcloud Talk video conferencing backend, recommended for calls with more than four participants. If you don't reserve a public IP, you can still use Talk in a more limited fashion (see the [Talk](#talk) section below for details).
+
+If you're not sure and just want the easiest, most affordable option, skip the public IP and use a gateway domain.
+
+
+
+# Deploy Nextcloud
+
+* On the [ThreeFold Dashboard](https://dashboard.grid.tf/), click on solutions from the sidebar, then click on **Nextcloud**
+* Choose a name for your deployment
+ * Note: You can use the auto-generated name if you want
+* Select a capacity package:
+ * **Minimum**: {cpu: 2, memory: 4gb, diskSize: 50gb }
+ * **Standard**: {cpu: 2, memory: 8gb, diskSize: 500gb }
+ * **Recommended**: {cpu: 4, memory: 16gb, diskSize: 1000gb }
+ * Or choose a **Custom** plan
+* If want to reserve a public IPv4 address, click on Network then select **Public IPv4**
+* If you want a [dedicated](../deploy/dedicated_machines.md) and/or a certified node, select the corresponding option
+* Choose the location of the node
+ * `Country`
+ * `Farm Name`
+* Select a node
+* If you want to use a custom domain, click on **Custom domain** under **Domain Name** and write your domain name
+ * Example: `nextcloudwebsite.com`
+* The **Select gateway** box will be visible whenever a gateway is required. If so, click it and choose a gateway
+ * If you are also using a custom domain, you must set your DNS record now before proceeding. The IP of the gateway will appear on screen. Check [below](#set-the-dns-record) for more information
+* Click **Deploy**
+
+
+
+# Nextcloud Setup
+
+Once the weblet is deployed, the details page will appear. If you are using a custom domain with a public IPv4, you'll need to set your DNS record now using the IP address shown under **Public IPv4**. Again, see [below](#dns-details) for details.
+
+Before you can access Nextcloud itself, you'll need to decide which addons you want to install and complete a setup step. This is done through the AIO interface that's included with your deployment. To access it, you can visit the **Nextcloud Setup** link shown in the details page, or click on the **Nextcloud Setup** button under **Actions** in the deployments list to set up Nextcloud.
+
+* Once you have access to the **Nextcloud AIO setup page**, you will be given a password composed of 8 words.
+ * Use this password to access the **Nextcloud AIO interface page**.
+ * Store this password somewhere safe. It's only possible to recover it by using SSH.
+* On the next page, you can add **Optionals addons** if you want.
+* Click on **Download and start containers** to start the Nextcloud instance.
+* Once the containers are properly started, you can access the Nextcloud admin login page by clicking **Open your Nextcloud**.
+ * You will be given an **Initial Nextcloud user name** and an **Initial Nextcloud password**. Use these credentials to log into the admin page.
+ * Store these credentials somewhere safe.
+* Later, if you want to access the Nextcloud admin login page, you can simply click on the button **Open Nextcloud** under **Actions** in the deployment list.
+
+The installation is now complete and you have access to your Nextcloud instance.
+
+
+
+# DNS Details
+
+## DNS Record with Public IPv4
+
+After deployment, you will have access to the IPv4 address of the VM you deployed on. You will need to add a **DNS A record** (Host: "@", Value: ) to your domain to access Nextcloud. This record type indicates the IP address of a given domain.
+
+## DNS Record with Gateway
+
+Before starting the deployment, you will need to add a **DNS A record** (Host: "@", Value: ) to your domain. The gateway IP will be shown to you when you select this option.
+
+## DNS Propagation
+
+When setting your own custom domain, it might take time for DNS to propagate. It is possible that you see the following message when opening the Nextcloud page:
+
+>"This site can't be reached. DNS address could not be found. Diagnosing the problem."
+
+This is normal. You might simply need to wait for the DNS to propagate completely.
+
+
+
+# Talk
+
+If you don't rent a public IP with your deployement, it's still possible to use Nextcloud Talk in a more limited fashion. It's generally understood that this method can work well for up to four participants in a call, and text chat also works without restriction. For larger calls, the dedicated backend, which requires a public IP, is recommended.
+
+While some calls can go entirely peer-to-peer and don't require any setup beyond installing the Talk app, a TURN server can be helpful to relay data when a peer-to-peer connection can't be established. There's more information on TURN servers after the install instructions.
+
+## Install Talk
+
+To install Talk, do the following:
+
+* Open the dropdown menu at the top right of the Nextcloud page
+* Click on **Apps**
+* In the left-side menu, select **Social & communication**
+* Scroll down and locate the Talk app
+* Click on **Download and enable**
+
+Once the Talk app is downloaded and enabled, you can find its icon at the top bar menu.
+
+## TURN
+
+As mentioned before, TURN servers relay data to help call participants connect to each other. All data sent to TURN server is encrypted in this case, so it's perfectly safe to use a free public server.
+
+That said, such free servers are not common, because relaying video chat uses a lot of bandwidth. As of the time of writing, Open Relay Project is one example that includes [instructions for use with Nextcloud Talk](https://www.metered.ca/tools/openrelay/#turn-server-for-nextcloud-talk).
+
+TURN server configuration can be found by opening the Talk settings, like this:
+
+* Open the dropdown menu at the top right of the Nextcloud page
+* Click on **Personal settings**
+* In the left-side menu, select **Talk**
+
+## Use Talk
+
+Once you've installed Talk and optionally added a TURN server, you can use Talk to create video conferences.
+
+Note that the host of the video meeting might need to turn the VPN off before creating a new conversation.
+
+
+
+# Backups and Updates
+
+## Create a Backup
+
+In the section **Backup and restore**, you can set a [BorgBackup](https://www.borgbackup.org/) of your Nextcloud instance.
+
+* Add a mount point and a directory name for your backup (e.g. **/mnt/backup**) and click **Submit backup location**.
+* After the creation of the backup location, write down the **encryption password for backups** somewhere safe and offline.
+* Click **Create backup** to create a BorgBackup of your Nextcloud instance.
+ * This will stop all containers, run the backup container and create the backup.
+* Once the backup is complete, you can click on **Start containers** to restart the Nextcloud instance.
+
+## Automatic Backups and Updates
+
+After the first manual backup of your Nextcloud instance is complete, you can set automatic backups and updates.
+
+* In the section **Backup and restore**, open the dropdown menu **Click here to reveal all backup options**.
+* In the section **Daily backup and automatic updates**, choose a time for your daily backup and click **Submit backup time**.
+ * To set automatic updates, make sure that the option **Automatically update all containers, the mastercontainer and on** is selected.
+
+
+
+# Troubleshooting
+
+## Retrieve the Nextcloud AIO Password
+
+You can retrieve the Nextcloud AIO password (8 words) by writing the following command line on the VM hosting your Nextcloud instance:
+
+```
+cat /mnt/data/docker/volumes/nextcloud_aio_mastercontainer/_data/data/configuration.json | grep password
+```
+
+## Access the Nextcloud Interface Page
+
+To access the Nextcloud interface page, follow those stepse
+
+* Open your Nextcloud instance
+* In the top right Profile menu, select **Administration Settings**
+* Under **Nextcloud All-in-One**, click **Open Nextcloud AIO Interface**
+
+
+
+## Check the DNS Propagation
+
+You can check if the DNS records are propagated globally with DNS propagation check services such as [DNS Checker](https://dnschecker.org/). You can use this tool to verify that your domain is properly pointing to the IPv4 address of the VM you deployed on.
+
+
+
+# Questions and Feedback
+
+If you have any questions, you can ask the ThreeFold community for help on the [ThreeFold Forum](http://forum.threefold.io/) or on the [ThreeFold Grid Tester Community](https://t.me/threefoldtesting) on Telegram.
\ No newline at end of file
diff --git a/collections/documentation/dashboard/solutions/nixos_micro.md b/collections/documentation/dashboard/solutions/nixos_micro.md
new file mode 100644
index 0000000..41f8932
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/nixos_micro.md
@@ -0,0 +1,66 @@
+ NixOS MicroVM
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Access the ThreeFold Dashboard](#access-the-threefold-dashboard)
+- [Deploy a NixOS MicroVM](#deploy-a-nixos-microvm)
+- [Questions and Feedback](#questions-and-feedback)
+
+***
+
+## Introduction
+
+__NixOS MicroVM__ refers to a minimalistic virtual machine environment based on the NixOS Linux distribution.
+The NixOS MicroVM leverages these principles to create a highly customizable and reproducible virtual machine environment. It allows users to define the entire system configuration, including packages, services, and dependencies, in a declarative manner using the Nix language. This ensures that the MicroVM is consistent, easily reproducible, and can be version-controlled.
+
+In this guide, will learn how to make reproducible, declarative and reliable systems by deploying a NixOS MicroVM weblet in ThreeFold Dashboard.
+
+For more information on Nix, you can read the [Nix Reference Manual](https://nixos.org/manual/nix/stable/).
+
+## Access the ThreeFold Dashboard
+
+* Go to the ThreeFold Dashboard website, based on the deployment network you prefer:
+ * [Mainnet](https://dashboard.grid.tf)
+ * [Testnet](https://dashboard.test.grid.tf)
+ * [Devnet](https://dashboard.dev.grid.tf)
+ * [QAnet](https://dashboard.qa.grid.tf)
+
+* Make sure you have a [wallet](../wallet_connector.md)
+* From the sidebar click on **Solutions**
+* Click on **Micro Virtual Machine** to start your NixOS MicroVM Deployment
+
+
+
+## Deploy a NixOS MicroVM
+
+We now present the main steps to properly configure your NixOS MicroVM running on the TFGrid.
+
+* In the section `Config`, make sure to select `Nixos` as the `VM Image`. You can choose different parameters (CPU, Memory, etc.) for your deployment depending on your workload needs.
+
+![](./img/nxios-micro1.png)
+
+* In the section `Environment Variables`, you can add the default configurations for Nix. Here's an example:
+ * ```
+ { pkgs ? import { } }:
+ let pythonEnv = pkgs.python3.withPackages(ps: [ ]); in pkgs.mkShell { packages = [ pythonEnv ]; }
+ ```
+ * This will be written to `/root/default.nix`. You can change the Nix shell configuration there.
+
+![](./img/nixos-micro2.png)
+
+* In the section `Disks`, you should mount a disk large enough for Nix to store its files used for `nix-store`.
+
+![](./img/nixos-micro3.png)
+
+* Once your configured the parameters, you can deploy the MicroVM.
+
+If you need more information on how to SSH into your deployment, read [this section](../../system_administrators/getstarted/tfgrid3_getstarted.md) of the TF Manual.
+
+
+
+## Questions and Feedback
+
+You should now be able to easily deploy a NixOS MicroVM on the ThreeFold Grid.
+
+If you have any question or feedback, you can write a post on the [ThreeFold Forum](http://forum.threefold.io/).
\ No newline at end of file
diff --git a/collections/documentation/dashboard/solutions/nodepilot.md b/collections/documentation/dashboard/solutions/nodepilot.md
new file mode 100644
index 0000000..1fc212a
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/nodepilot.md
@@ -0,0 +1,52 @@
+ NodePilot
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Prerequisites](#prerequisites)
+- [Deployment](#deployment)
+
+***
+
+## Introduction
+
+This is a simple instance of upstream [Node Pilot](https://nodepilot.tech).
+
+## Prerequisites
+
+- Make sure you have a [wallet](../wallet_connector.md)
+- From the sidebar click on **Applications**
+- Click on **Node Pilot**
+
+## Deployment
+
+![ ](./img/solutions_nodepilot.png)
+
+- Fill in the instance name: it's used to reference the node-pilot in the future.
+
+- Minimum CPU allowed is 8 cores and minimum memory allowed is 8192.
+
+- `Dedicated` flag to retrieve only dedeicated nodes
+- `Certified` flag to retrieve only certified nodes
+
+- Choose the location of the node
+ - `Region`
+ - `Country`
+ - `Farm Name`
+- Select a node to deploy your node-pilot instance on.
+
+> Or you can select a specific node with manual selection.
+
+- When using the [flist](https://hub.grid.tf/tf-official-vms/node-pilot-zdbfs.flist) you get a node pilot instance ready out-of-box. You need to get a public ipv4 to get it to works.
+
+After that is done you can see a list of all of your deployed instances
+
+![ ](./img/nodeP_2.png)
+
+Click on ***Visit*** to go to the registration page of your Node Pilot instance!
+
+![ ](./img/nodePilot_3.png)
+
+You can go to `https://publicip` and configure your node-pilot. You can upload a backup to the VM via ssh as well if you have a backup of a previous instance.
+
+What change compared to upstream node-pilot, we have out-of-box a transparent pre-filled blockchain database for some blochain (currently Fuse and Pokt as proof-of-concept). You can start one of theses blockchain in no-time and it will be automatically nearly sync already without the requirement of the full space locally nor downloading everything and killing bandwidth.
diff --git a/collections/documentation/dashboard/solutions/owncloud.md b/collections/documentation/dashboard/solutions/owncloud.md
new file mode 100644
index 0000000..16726a0
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/owncloud.md
@@ -0,0 +1,86 @@
+ ownCloud
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Prerequisites](#prerequisites)
+- [Deploy ownCloud](#deploy-owncloud)
+ - [Base](#base)
+ - [SMTP](#smtp)
+ - [List of Instances](#list-of-instances)
+- [Admin Connection](#admin-connection)
+- [TFConnect App Connection](#tfconnect-app-connection)
+
+***
+
+## Introduction
+
+[ownCloud](https://owncloud.com/) develops and provides open-source software for content collaboration, allowing teams to easily share and work on files seamlessly regardless of device or location.
+
+## Prerequisites
+
+- Make sure you have a [wallet](../wallet_connector.md)
+- From the sidebar click on **Solutions**
+- Click on **ownCloud**
+
+## Deploy ownCloud
+
+![ ](./img/owncloud1.png)
+
+### Base
+
+- Enter an ownCloud deployment name.
+ - The name is used in generating a unique subdomain on one of the gateways on the network alongside your twin ID.
+ - Ex. ***oc98newcloud*.gent02.dev.grid.tf**
+
+- Enter administrator information including **Username** and **Password**.
+ - This admin user will have full permission on the deployed instance.
+- Select a capacity package:
+ - **Small**: {cpu: 2, memory: 8, diskSize: 250 }
+ - **Medium**: {cpu: 2, memory: 16, diskSize: 500 }
+ - **Large**: {cpu: 4, memory: 32, diskSize: 1000 }
+ - Or choose a **Custom** plan
+- Choose the network
+ - `Public IPv4` flag gives the virtual machine a Public IPv4
+- Enable the `Dedicated` flag to retrieve only dedicated nodes
+- Enable the `Certified` flag to retrieve only certified nodes
+- Choose the location of the node
+ - `Region`
+ - `Country`
+ - `Farm Name`
+- Choose the node to deploy on
+> Or you can select a specific node with manual selection.
+- Enable the `Custom Domain` flag to use a custom domain
+- Choose a gateway node to deploy your ownCloud instance on.
+
+Once you've set the deployment parameters, you can click on **Deploy**.
+
+### SMTP
+
+On the SMTP window, you can enable the optional `SMTP Server` flag if you want to have your ownCloud instance configured with an SMTP server.
+
+![ ](./img/owncloud4.png)
+
+### List of Instances
+
+When the deployment is ready, you will see a list of all of your deployed instances.
+
+![ ](./img/owncloud5.png)
+
+## Admin Connection
+
+Click on the button **Visit** under **Actions** to open the ownCloud login window. If you see **bad gateway**, you might simply need to wait a couple of minutes until the deployment completes.
+
+![ ](./img/solutions_owncloud_visit.png)
+
+To consult the deployment details, click on the button **Details** under **Actions**. On this page, you can access the **ownCloud Admin Username** and the **ownCloud Admin Password**. Use those credentials to log in as an administrator on your ownCloud deployment.
+
+![ ](./img/owncloud6.png)
+
+## TFConnect App Connection
+
+To connect to your ownCloud instance with the ThreeFold Connect app, you need to add permissions to your ThreeFold 3Bot ID by first [connecting as an administrator](#admin-connection).
+
+- Once you're connected as an admin, open the top-right menu and click on **Users**.
+- To create a new user, set your 3Bot ID as the username with its corresponding email address, and set **Groups** as **admin**. Then click **Create**.
+- You can now log out and connect to your ownCloud instance with the TF Connect app.
\ No newline at end of file
diff --git a/collections/documentation/dashboard/solutions/peertube.md b/collections/documentation/dashboard/solutions/peertube.md
new file mode 100644
index 0000000..79a74ee
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/peertube.md
@@ -0,0 +1,59 @@
+ Peertube
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Prerequisites](#prerequisites)
+- [Deployment](#deployment)
+
+***
+
+## Introduction
+
+[Peertube](https://joinpeertube.org/) aspires to be a decentralized and free/libre alternative to video broadcasting services.
+
+## Prerequisites
+
+- Make sure you have a [wallet](../wallet_connector.md)
+- From the sidebar click on **Applications**
+- Click on **Peertube**
+
+## Deployment
+
+![ ](./img/solutions_peertube.png)
+
+- Enter an Application Name. It's used in generating a unique subdomain on one of the gateways on the network alongside your twin ID.
+ the applied format `.` Ex. ***pt100peerprod*.gent02.dev.grid.tf**
+- Enter an email and password which will be used for the admin login.
+- Select a capacity package:
+ - **Small**: { cpu: 1, memory: 2, diskSize: 15 }
+ - **Medium**: { cpu: 2, memory: 4, diskSize: 100 }
+ - **Large**: { cpu: 4, memory: 16, diskSize: 250 }
+ - Or choose a **Custom** plan
+
+ - `Public IPv4` flag gives the virtual machine a Public IPv4
+ - `Public IPv6` flag gives the virtual machine a Public IPv6
+ - `Planetary Network` to connect the Virtual Machine to Planetary network
+ - `Wiregaurd Access` to add a wiregaurd acces to the Virtual Machine
+- `Dedicated` flag to retrieve only dedeicated nodes
+- `Certified` flag to retrieve only certified nodes
+- Choose the location of the node
+ - `Region`
+ - `Country`
+ - `Farm Name`
+
+- Choose the node to deploy on
+> Or you can select a specific node with manual selection.
+- `Custom Domain` flag lets the user to use a custom domain
+- Choose a gateway node to deploy your Peertube instance on.
+
+After that is done you can see a list of all of your deployed instances
+
+
+![ ](./img/weblet_peertube_listing.png)
+
+Click on ***Visit*** to go to the homepage of your Peertube instance!
+
+![ ](./img/weblet_peertube_instance.png)
+
+> Please note it may take sometime to be ready
\ No newline at end of file
diff --git a/collections/documentation/dashboard/solutions/presearch.md b/collections/documentation/dashboard/solutions/presearch.md
new file mode 100644
index 0000000..c0c3f55
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/presearch.md
@@ -0,0 +1,98 @@
+ Presearch
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Prerequisites](#prerequisites)
+- [Deploy a Presearch Node](#deploy-a-presearch-node)
+- [Migrate an Existing Presearch Node to the TFGrid](#migrate-an-existing-presearch-node-to-the-tfgrid)
+- [Verify if a 3Node Already Runs a Presearch Workload](#verify-if-a-3node-already-runs-a-presearch-workload)
+- [Learn More About Presearch](#learn-more-about-presearch)
+- [Questions and Feedback](#questions-and-feedback)
+
+***
+
+## Introduction
+
+[Presearch](https://www.presearch.io/) is a community-powered, decentralized search engine that provides better results while protecting your privacy and rewarding you when you search. This weblet deploys a Presearch node. Presearch Nodes are used to process user search requests, and node operators earn Presearch PRE tokens for joining and supporting the network.
+
+## Prerequisites
+
+- Make sure you have a [wallet](../wallet_connector.md)
+- From the sidebar click on **Applications**
+- Click on **Presearch**
+
+## Deploy a Presearch Node
+
+![ ](./img/solutions_presearch.png)
+
+- Enter an instance name.
+
+- You need to sign up on Presearch in order to get your *Presearch Registration Code*. To sign up, go to the [Presearch](https://presearch.com) website, create your account and then head to your [dashboard](https://nodes.presearch.com/dashboard) to find your registration code.
+
+- Choose the network
+ - `Public IPv4` flag gives the virtual machine a Public IPv4
+ - `Planetary Network` to connect the Virtual Machine to Planetary network
+
+- `Dedicated` flag to retrieve only dedeicated nodes
+- `Certified` flag to retrieve only certified nodes
+- Choose the location of the node
+ - `Region`
+ - `Country`
+ - `Farm Name`
+
+- Choose the node to deploy the Virtual Machine on
+> Or you can select a specific node with manual selection.
+
+## Migrate an Existing Presearch Node to the TFGrid
+
+Now what if you already have a Presearch node deployed somewhere and would like to migrate to Threefold?
+
+We got you! All you need to do is:
+
+1. Login to your old server that has your node via SSH.
+2. Run `docker cp presearch-node:/app/node/.keys presearch-node-keys` in order to generate your key-pair.
+3. Head to the *Restore* tab in the Presearch weblet and paste your key-pair in the fields below and you'll be good to deploy!
+
+![ ](./img/presearch6.png)
+
+After that is done you can see a list of all of your deployed instances
+
+![ ](./img/presearch4.png )
+
+Now head to your [dashboard](https://nodes.presearch.com/dashboard) again and scroll down to **Current Nodes**, you'll see your newly created node up and connected!
+
+![ ](./img/presearch5.png)
+
+
+## Verify if a 3Node Already Runs a Presearch Workload
+
+You can do the following to verify if a Presearch workload deployed without a public IP address already has a Presearch workload running. Note that you will first need to deploy a Presearch workload on the 3Node. After deployment, you can SSH into the VM and do the verification.
+
+* SSH into the VM running the Presearch workload
+ * ```
+ ssh root@
+ ```
+* List running containers and identity the Presearch container
+ * ```
+ docker ps
+ ```
+* Print the logs of the Presearch container
+ * ```
+ docker logs
+ ```
+* If there is no other Presearch workload running on the 3Node, you will see a similar output:
+ * > 2023-10-16T12:18:33.780Z info: Node is listening for searches...
+* If there is another Presearch workload running on the 3Node, you will see a similar output:
+ * > 2023-10-16T12:24:00.346Z error: Duplicate IP: This IP Address is already running another Node. Only one Node is permitted per IP Address.
+ * If there is another Presearch workload running, you will need to either deploy on another 3Node with a public IP or deploy on another node without a public IP that isn't running a Presearch deployment.
+
+
+
+## Learn More About Presearch
+
+To learn more about Presearch, you can read the [Presearch documentation](https://docs.presearch.io/).
+
+## Questions and Feedback
+
+If you have any questions, you can ask the ThreeFold community for help on the [ThreeFold Forum](http://forum.threefold.io/) or on the [ThreeFold Grid Tester Community](https://t.me/threefoldtesting) on Telegram.
\ No newline at end of file
diff --git a/collections/documentation/dashboard/solutions/solutions.md b/collections/documentation/dashboard/solutions/solutions.md
new file mode 100644
index 0000000..60732e3
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/solutions.md
@@ -0,0 +1,29 @@
+Solutions
+
+This section provides a non-code easy way to deploy a whole solution on the TFGrid.
+
+Table of Contents
+
+- [Basic Environments](./basic_environments_readme.md)
+ - [Virtual Machines](./vm_intro.md)
+ - [Micro and Full VM Differences](./vm_differences.md)
+ - [Full Virtual Machine](./fullVm.md)
+ - [Micro Virtual Machine](./vm.md)
+ - [Kubernetes](./k8s.md)
+ - [NixOS MicroVM](./nixos_micro.md)
+- [Ready Community Solutions](./ready_community_readme.md)
+ - [Caprover](./caprover.md)
+ - [Funkwhale](./funkwhale.md)
+ - [Peertube](./peertube.md)
+ - [Taiga](./taiga.md)
+ - [Owncloud](./owncloud.md)
+ - [Nextcloud](./nextcloud.md)
+ - [Discourse](./discourse.md)
+ - [Mattermost](./mattermost.md)
+ - [Presearch](./presearch.md)
+ - [CasperLabs](./casper.md)
+ - [Node Pilot](./nodepilot.md)
+ - [Subsquid](./subsquid.md)
+ - [Algorand](./algorand.md)
+ - [Wordpress](./wordpress.md)
+ - [Umbrel](./umbrel.md)
diff --git a/collections/documentation/dashboard/solutions/subsquid.md b/collections/documentation/dashboard/solutions/subsquid.md
new file mode 100644
index 0000000..063ff35
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/subsquid.md
@@ -0,0 +1,54 @@
+ Subsquid
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Prerequisites](#prerequisites)
+- [Deployment](#deployment)
+
+***
+
+## Introduction
+
+[Subsquid](https://www.subsquid.io/) indexer is a piece of software that reads all the blocks from a Substrate based blockchain, decodes and stores them for processing in a later stage.
+
+## Prerequisites
+
+- Make sure you have a [wallet](../wallet_connector.md)
+- From the sidebar click on **Applications**
+- Click on **Subsquid**
+
+## Deployment
+
+![ ](./img/solutions_subsquid.png)
+
+- Enter an instance name.
+
+- Enter an endpoint for a supported substrate chain. You can find the list of endpoints of supported cahins [here](https://github.com/polkadot-js/apps/blob/master/packages/apps-config/src/endpoints/production.ts).
+
+
+- Select a capacity package:
+ - **Small**: {cpu: 1, memory: 2 , diskSize: 50 }
+ - **Medium**: {cpu: 2, memory: 4, diskSize: 100 }
+ - **Large**: {cpu: 4, memory: 16, diskSize: 250 }
+ - Or choose a **Custom** plan
+
+- `Dedicated` flag to retrieve only dedeicated nodes
+- `Certified` flag to retrieve only certified nodes
+- Choose the location of the node
+ - `Region`
+ - `Country`
+ - `Farm Name`
+- Choose the node to deploy on
+> Or you can select a specific node with manual selection.
+- `Custom Domain` flag lets the user to use a custom domain
+- Choose a gateway node to deploy your Subsquid instance on.
+
+
+After that is done you can see a list of all of your deployed instances
+
+![ ](./img/subsquid_list.png)
+
+Click on ***Visit*** to go to the homepage of your Subsquid indexer instance!
+
+![ ](./img/subsquid_graphql.png)
diff --git a/collections/documentation/dashboard/solutions/taiga.md b/collections/documentation/dashboard/solutions/taiga.md
new file mode 100644
index 0000000..4291e12
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/taiga.md
@@ -0,0 +1,57 @@
+ Taiga
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Prerequisites](#prerequisites)
+- [Deployment](#deployment)
+
+***
+
+## Introduction
+
+[Taiga](https://www.taiga.io/) is the project management tool for multi-functional agile teams. It has a rich feature set and at the same time it is very simple to start with through its intuitive user interface.
+
+## Prerequisites
+
+- Make sure you have a [wallet](../wallet_connector.md)
+- From the sidebar click on **Applications**
+- Click on **Taiga**
+
+## Deployment
+
+![ ](./img/solutions_taiga.png)
+
+- Enter an Application Name. It's used in generating a unique subdomain on one of the gateways on the network alongside your twin ID. Ex. ***tg98taigar*.gent02.dev.grid.tf**
+
+- Enter administrator information including **Username**, **Email** and **Password**. This admin user will have full permission on the deployed instance.
+- Select a capacity package:
+ - **Small**: {cpu: 2, memory: 4, diskSize: 100 }
+ - **Medium**: {cpu: 4, memory: 8, diskSize: 150 }
+ - **Large**: {cpu: 4, memory: 16, diskSize: 250 }
+ - Or choose a **Custom** plan
+
+- `Dedicated` flag to retrieve only dedeicated nodes
+- `Certified` flag to retrieve only certified nodes
+- Choose the location of the node
+ - `Region`
+ - `Country`
+ - `Farm Name`
+- Choose the node to deploy the Tiaga instance on
+> Or you can select a specific node with manual selection.
+- `Custom Domain` flag lets the user to use a custom domain
+- Choose a gateway node to deploy your Funkwhale instance on.
+
+
+
+There's also an optional **Mail Server** tab if you'd like to have your Taiga instance configured with an SMTP server.
+
+![ ](./img/taiga4.png)
+
+After that is done you can see a list of all of your deployed instances
+
+![ ](./img/taiga5.png)
+
+Click on ***Visit*** to go to the homepage of your Taiga instance!
+
+![ ](./img/taiga6.png)
diff --git a/collections/documentation/dashboard/solutions/umbrel.md b/collections/documentation/dashboard/solutions/umbrel.md
new file mode 100644
index 0000000..21075ca
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/umbrel.md
@@ -0,0 +1,60 @@
+ Umbrel
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Prerequisites](#prerequisites)
+
+***
+
+## Introduction
+
+[Umbrel](https://umbrel.com/) is an OS for running a personal server in your home. Self-host open source apps like Nextcloud, Bitcoin node, and more.
+
+## Prerequisites
+
+- Make sure you have a [wallet](../wallet_connector.md)
+- From the sidebar click on **Applications**
+- Click on **Umbrel**
+
+**Process** :
+![Config](./img/solutions_umbrel.png)
+
+- Enter an instance name.
+- Enter a Username
+ - will be used to create Umbrel dashboard account.
+- Enter a Password
+ - Will be used to login to the Umbrel dashboard.
+ - Must be 12 to 30 characters .
+- Select a capacity package:
+ - **Small**: { cpu: 1, memory: 2, diskSize: 10 }
+ - **Medium**: { cpu: 2, memory: 4 , diskSize: 50 }
+ - **Large**: { cpu: 4, memory: 16 , diskSize: 100 }
+ - Or choose a **Custom** plan
+
+- `Dedicated` flag to retrieve only dedeicated nodes
+- `Certified` flag to retrieve only certified nodes
+- Choose the location of the node
+ - `Region`
+ - `Country`
+ - `Farm Name`
+- Choose the node to deploy the Umbrel instance on
+> Or you can select a specific node with manual selection.
+
+**After Deploying**:
+
+You can see a list of all of your deployed instances
+
+![ ](./img/umbrel2.png)
+
+- you can click on `Show details` for more details about the Umbrel deployment.
+ ![ ](./img/umbrel3.png)
+ and for more detailed information switch to `JSON` tap.
+ ![ ](./img/umbrel4.png)
+- Click on ***Admin Panel*** to go to the dashboard of your Umbrel instance!
+ - Enter the ***Password*** that you provided in `config` section to login into Umbrel dashboard.
+ > Forget the credentials? You can find them with `Show details` button.
+
+![ ](./img/umbrel5.png)
+
+> **Warning**: Due to the nature of the grid, shutdown, or restart your umbrel from the dashboard **MAY** make some unwanted behaviors.
diff --git a/collections/documentation/dashboard/solutions/vm.md b/collections/documentation/dashboard/solutions/vm.md
new file mode 100644
index 0000000..5e8db2f
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/vm.md
@@ -0,0 +1,62 @@
+ Micro Virtual Machine
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Deployment](#deployment)
+
+***
+
+## Introduction
+
+We present the steps to deploy a micro VM on the TFGrid.
+
+
+## Deployment
+
+Deploy a new virtual machine on the Threefold Grid
+
+- Make sure you have a [wallet](../wallet_connector.md)
+- From the sidebar click on **Solutions**
+- Click on **Micro Virtual Machine**
+
+__Process__ :
+
+![ ](./img/solutions_microvm.png)
+
+- Fill in the instance name: it's used to reference the VM in the future.
+- Choose the image from the drop down (e.g Alpine, Ubuntu) or you can click on `Other` and manually specify the flist URL and the entrypoint.
+- Select a capacity package:
+ - **Small**: {cpu: 1, memory: 2, diskSize: 25 }
+ - **Medium**: {cpu: 2, memory: 4, diskSize: 50 }
+ - **Large**: {cpu: 4, memory: 16, diskSize: 100}
+ - Or choose a **Custom** plan
+- Choose the network
+ - `Public IPv4` flag gives the virtual machine a Public IPv4
+ - `Public IPv6` flag gives the virtual machine a Public IPv6
+ - `Planetary Network` to connect the Virtual Machine to Planetary network
+ - `Wireguard Access` to add a wireguard acces to the Virtual Machine
+- `GPU` flag to add GPU to the Virtual machine
+- `Dedicated` flag to retrieve only dedicated nodes
+- `Certified` flag to retrieve only certified nodes
+- Choose the location of the node
+ - `Region`
+ - `Country`
+ - `Farm Name`
+- Choose the node to deploy the Virtual Machine on
+> Or you can select a specific node with manual selection.
+
+![](./img/nixos-micro2.png)
+* In the section `Environment Variables`, you can add any environment variables that the machine might need
+
+![](./img/nixos-micro3.png)
+
+* In the section `Disks`, You can attach one or more disks to the Virtual Machine by clicking on the Disks tab and the plus `+` sign and specify the following parameters
+ - Disk name
+ - Disk size
+
+in the bottom of the page you can see a list of all of the virual machines you deployed. you can click on `Show details` for more details
+
+![](./img/vm_list.png)
+You can also go to JSON tab for full details
+![ ](./img/vm_json.png)
\ No newline at end of file
diff --git a/collections/documentation/dashboard/solutions/vm_differences.md b/collections/documentation/dashboard/solutions/vm_differences.md
new file mode 100644
index 0000000..a6dff11
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/vm_differences.md
@@ -0,0 +1,29 @@
+ Micro and Full VM Differences
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Micro Virtual Machine](#micro-virtual-machine)
+- [Full Virtual Machine](#full-virtual-machine)
+
+***
+
+## Introduction
+
+We present the main differences between a micro VM and a full VM. This is useful information when it comes to choosing the proper deployment on the TFGrid.
+
+## Micro Virtual Machine
+
+- It's meant to host microservice. and the user should enter the entrypoint.
+- The user has no control over ther kernel used to run the machine.
+- The network setup will be created for the user. And the vm's init process can assume that it will be fully set up (according to the config the user provided) by the time it is started.
+- Mountpoints will also be setup for the user. The environment variables passed will be available inside the the vm.
+
+## Full Virtual Machine
+
+- The users run their own operating system, but the image must be
+ - EFI bootable
+ - Cloud-init enabled
+- It contains a default disk attached, as the boot image will be copied to this disk.
+- The default disk is mounted on / so if you want to attach any additional disks, you have to choose a different mounting point.
+- A /image.raw file is used as "boot disk". This /image.raw is copied to the first attached volume of the vm. Cloud-init will take care of resizing the filesystem on the image to take the full disk size allocated in the deployment.
\ No newline at end of file
diff --git a/collections/documentation/dashboard/solutions/vm_intro.md b/collections/documentation/dashboard/solutions/vm_intro.md
new file mode 100644
index 0000000..d9cc0a7
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/vm_intro.md
@@ -0,0 +1,11 @@
+ Virtual Machines
+
+On the TFGrid, you can deploy both micro and full virtual machines.
+
+ Table of Contents
+
+- [Micro and Full VM Differences ](./vm_differences.md)
+- [Full Virtual Machine](./fullVm.md)
+- [Micro Virtual Machine](./vm.md)
+- [Nixos MicroVM](./nixos_micro.md)
+- [Add a Domain](./add_domain.md)
\ No newline at end of file
diff --git a/collections/documentation/dashboard/solutions/wordpress.md b/collections/documentation/dashboard/solutions/wordpress.md
new file mode 100644
index 0000000..57f2938
--- /dev/null
+++ b/collections/documentation/dashboard/solutions/wordpress.md
@@ -0,0 +1,147 @@
+ WordPress
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Prerequisites](#prerequisites)
+- [Domain Name and IP Address](#domain-name-and-ip-address)
+- [DNS Details with Custom Domain](#dns-details-with-custom-domain)
+ - [DNS Record with Public IPv4](#dns-record-with-public-ipv4)
+ - [DNS Record with Gateway](#dns-record-with-gateway)
+ - [DNS Propagation](#dns-propagation)
+- [Deployment Process](#deployment-process)
+- [Access WordPress](#access-wordpress)
+ - [WordPress Instance Website](#wordpress-instance-website)
+ - [WordPress Instance Admin Page](#wordpress-instance-admin-page)
+- [WordPress Instance Credentials](#wordpress-instance-credentials)
+- [Questions and Feedback](#questions-and-feedback)
+
+***
+
+# Introduction
+
+[WordPress](https://wordpress.org/) is the most popular CMS on the market, powering 65.2% of websites whose CMS we know. That translates to 42.4% of all websites – nearly half of the internet. It is a popular option for those who want to build a website or a blog.
+
+# Prerequisites
+
+- Make sure you have a [wallet](../wallet_connector.md)
+- From the sidebar click on **Solutions**
+- Click on **Wordpress**
+
+# Domain Name and IP Address
+
+A domain name is required to use WordPress. You can either use your own, which we'll call a **custom domain**, or you can get a free subdomain from a gateway node. Note that this won't impact the function of your deployment, it's just a matter of preference. If you want to use your own domain, follow the steps for custom domain wherever you see them below.
+
+Another choice to make before launching your WordPress instance is whether you want to reserve a public IPv4 for the deployment. Note that renting a public IPv4 address is an extra cost. If you do not enable IPv4, the deployment will be provided a gateway IPv4 address.
+
+If you're not sure and just want the easiest, most affordable option, do not enable public IPv4 nor custom domain.
+
+# DNS Details with Custom Domain
+
+In this section, we cover the essential DNS information when deploying a WordPress instance with a custom domain.
+
+You can skip this section if you did not enable **Custom Domain** in **Domain Name**.
+
+As a general reference, here is what setting a DNS A record can look like:
+
+![ ](img/wp11.png)
+
+This record type indicates the IP address of a given domain.
+
+## DNS Record with Public IPv4
+
+Consider the following if you've enabled **Custom Domain** in **Domain Name** and **Public IPv4** in **Network**.
+
+After deployment, you will have access to the IPv4 address of the VM you deployed on. You will need to add a **DNS A record** (Host: "@", Value: ) to your domain to access WordPress.
+
+## DNS Record with Gateway
+
+Consider the following if you've enabled **Custom domain** in **Domain Name** but did not enable **Public IPv4** in **Network**.
+
+Before deploying the WordPress instance, you will have access to the gateway IPv4 address. You will need to add a **DNS A record** (Host: "@", Value: ) to your domain to access WordPress.
+
+## DNS Propagation
+
+When setting a DNS A record, it might take time for the DNS to propagate. It is possible that you see the following message when opening the WordPress page:
+
+>"This site can't be reached. DNS address could not be found. Diagnosing the problem."
+
+This is normal. You might simply need to wait for the DNS to propagate completely.
+
+You can check if the DNS records are propagated globally with DNS propagation check services such as [DNS Checker](https://dnschecker.org/). You can use this tool to verify that your domain is properly pointing to either the VM or the gateway IPv4 address.
+
+# Deployment Process
+
+In this section, we cover the steps to deploy a WordPress instance on the Playground.
+
+![Config](./img/solutions_wordpress.png)
+
+- Enter an instance name or leave the auto-generated instance name
+- Enter the admin information or leave the auto-generated information
+ - **Username**: This will be used as the MySQL DB username and for Wp-admin
+ - **Password**: This will be used as the MySQL DB password and for Wp-admin
+ - **Email**: This will be used for Wp-admin
+- Select a capacity package:
+ - **Small**: { cpu: 1, memory: 2 , diskSize: 15 }
+ - **Medium**: { cpu: 2, memory: 4 , diskSize: 50 }
+ - **Large**: { cpu: 4, memory: 16 , diskSize: 100 }
+ - Or choose a **Custom** plan
+
+- Choose the network
+ - **Public IPv4** flag gives the virtual machine a Public IPv4
+
+- **Dedicated** flag to retrieve only dedicated nodes
+- **Certified** flag to retrieve only certified nodes
+- Choose the location of the node
+ - **Country**
+ - **Farm Name**
+- Choose the node to deploy the WordPress instance on
+- **Custom Domain** flag lets the user to use a custom domain
+- Choose a gateway node to deploy your WordPress instance on
+ - If you've enabled IPv4, you do not need to choose a gateway node
+
+# Access WordPress
+
+In the section **WordPress Instances**, you can see a list of all of your deployed instances:
+
+![ ](./img/wp2.png)
+
+You can click on **Show details** under **Actions** for more details about the WordPress deployment.
+
+![ ](img/wp8.png)
+
+![ ](img/wp3.png)
+
+For more detailed information, you can switch to the **Json** tab.
+
+![ ](img/wp4.png)
+
+## WordPress Instance Website
+
+Click on **Visit** under **Actions** to go to the homepage of your WordPress instance.
+
+![ ](img/wp10.png)
+
+![ ](img/wp5.png)
+
+## WordPress Instance Admin Page
+
+Click on **Admin Panel** to go to the WordPress admin page (**wp-admin**) of your WordPress instance.
+
+![ ](img/wp9.png)
+
+Enter the **Username** and the **Password** that you provided in the **config** section to log into the admin panel.
+
+![ ](img/wp6.png)
+
+![ ](img/wp7.png)
+
+# WordPress Instance Credentials
+
+At any time, you can find the credentials of your WordPress instance by clicking on the **Show details** button under **Actions**.
+
+![ ](img/wp8.png)
+
+# Questions and Feedback
+
+If you have any questions, you can ask the ThreeFold community for help on the [ThreeFold Forum](http://forum.threefold.io/) or on the [ThreeFold Grid Tester Community](https://t.me/threefoldtesting) on Telegram.
diff --git a/collections/documentation/dashboard/tfchain/tf_dao.md b/collections/documentation/dashboard/tfchain/tf_dao.md
new file mode 100644
index 0000000..648280b
--- /dev/null
+++ b/collections/documentation/dashboard/tfchain/tf_dao.md
@@ -0,0 +1,41 @@
+DAO Voting
+
+The TFChain DAO (i.e. Decentralized Autonomous Organization) feature integrates decentralized governance capabilities into the ThreeFold Dashboard. It enables community members to participate in decision-making processes and to contribute to the evolution of the ThreeFold ecosystem. Through the TFChain DAO, users can propose, vote on, and implement changes to the network protocols, policies, and operations, fostering a collaborative and inclusive environment.
+
+Table of Contents
+
+- [An Introduction to the DAO concept](#an-introduction-to-the-dao-concept)
+- [Prerequisites to Vote](#prerequisites-to-vote)
+- [How to Vote for a Proposal](#how-to-vote-for-a-proposal)
+- [The Goal of the Threefold DAO](#the-goal-of-the-threefold-dao)
+
+***
+
+## An Introduction to the DAO concept
+
+[A decentralized autonomous organization (DAO)](../../../knowledge_base/about/dao/dao.md) is an entity with no central leadership. Decisions get made from the bottom-up, governed by a community organized around a specific set of rules enforced on a blockchain.
+
+DAOs are internet-native organizations collectively owned and managed by their members. They have built-in treasuries that are only accessible with the approval of their members. Decisions are made via proposals the group votes on during a specified period.
+
+
+
+## Prerequisites to Vote
+
+Voting for a DAO proposal is very simple. You first need to meet certain requirements to be able to vote.
+
+- Have a [Threefold farm](../farms/farms.md)
+- Have at least one active [3node server](../../farmers/3node_building/3node_building.md) on the farm
+- If you created your farm with the Threefold Connect app
+ - [Import your farm on the Threefold Dashboard](../../threefold_token/storing_tft/tf_connect_app.md#move-farm-from-the-tf-connect-app-to-the-tf-portal-polkadotjs)
+
+
+
+## How to Vote for a Proposal
+
+To vote, you need to log into your Threefold Dashboard account, go to **TF DAO** section of **TFChain** and vote for an active proposal. Make sure to read the proposition and ask questions on the Threefold Forum proposition post if you have any.
+
+## The Goal of the Threefold DAO
+
+The goal of DAO voting system is to gather the thoughts and will of the Threefold community and build projects that are aligned with the ethos of the project.
+
+We encourage anyone to share their ideas. Who knows? Your sudden spark of genius might lead to an accepted proposal on the Threefold DAO!
diff --git a/collections/documentation/dashboard/tfchain/tf_minting_reports.md b/collections/documentation/dashboard/tfchain/tf_minting_reports.md
new file mode 100644
index 0000000..158ba04
--- /dev/null
+++ b/collections/documentation/dashboard/tfchain/tf_minting_reports.md
@@ -0,0 +1,5 @@
+# TF Minting Reports
+
+TFGrid Minting Explorer, works by simply entering a receipt hash and you will get full minting report in return.
+
+![](../img/Minting.png)
\ No newline at end of file
diff --git a/collections/documentation/dashboard/tfchain/tf_token_bridge.md b/collections/documentation/dashboard/tfchain/tf_token_bridge.md
new file mode 100644
index 0000000..f7391c2
--- /dev/null
+++ b/collections/documentation/dashboard/tfchain/tf_token_bridge.md
@@ -0,0 +1,44 @@
+# TF Token Bridge
+
+Transferring TFT between Stellar and Tfchain
+
+## Usage
+
+This document will explain how you can transfer TFT from Tfchain to Stellar and back.
+
+![bridge page](../img/bridge.png)
+
+## Prerequisites
+
+- Stellar wallet
+
+- Account on TFchain (use TF Dashboard to create one).
+
+## Stellar to Tfchain
+
+You can deposit to Tfchain using the bridge page on the TF Dashboard
+
+A deposit of tokens from the Stellar network onto TF-Chain needs to happen from a Stellar wallet, like in the ThreeFold Connect App.
+
+You have 2 options:
+
+- TFT needs to be sent to the bridge account
+- specifying in the memo field the twinID that was generated with the Twin creation e.g. twin_110 (dont forget twin_)
+- Note there is a transaction cost of 1 TFT.
+
+Or
+
+- You can scan the QR code
+
+![bridge](../img/bridge_deposit.png)
+
+## Tfchain to Stellar
+
+You can bridge back to stellar using the bridge page on the dashboard, click *withdraw*:
+
+After indicating the destination address and the amount to be transferred, click *Send*.
+
+![withdraw](../img/bridge_withdraw.png)
+
+A withdraw fee of 1 TFT will be taken, so make sure you send a larger amount as 1 TFT.
+The amount withdrawn from TFChain will be sent to your Stellar wallet.
diff --git a/collections/documentation/dashboard/tfchain/tf_token_transfer.md b/collections/documentation/dashboard/tfchain/tf_token_transfer.md
new file mode 100644
index 0000000..733eff6
--- /dev/null
+++ b/collections/documentation/dashboard/tfchain/tf_token_transfer.md
@@ -0,0 +1,26 @@
+# TF Token Transfer
+
+Manage your TFT on TFChain.
+
+## Transfer TFT between TFChain accounts
+
+You can transfer TFTs between two accounts that exist on the same chain.
+
+> Remark: testnet and mainnet both have the same TFTs but as the 2 chains are different, there is no way to do a direct transfer between accounts on testnet and on mainnet.
+
+You can transfer TFTs by recipient address or recipient twin id; by selecting the needed tab.
+
+
+### Transfer by twin id
+
+Fill in the recipient twin id, the amount of tokens to transfer, and click on `Send`.
+
+![](../img/dashboard_transfer_twin.png)
+
+### Transfer by address
+
+Fill in the recipient address, the amount of tokens to transfer, and click on `Send`.
+
+![](../img/dashboard_transfer_address.png)
+
+There is no transfer fee, just a signing fee of `0.001` TFT.
diff --git a/collections/documentation/dashboard/tfchain/tfchain.md b/collections/documentation/dashboard/tfchain/tfchain.md
new file mode 100644
index 0000000..bda6640
--- /dev/null
+++ b/collections/documentation/dashboard/tfchain/tfchain.md
@@ -0,0 +1,20 @@
+# TFChain
+
+Here you will find everything related to the ThreeFold chain. this includes:
+
+- Detailed account information from the [Your Profile](./your_profile.md) section.
+- Information about what DAO is and how to vote on DAO proposals from the [TF DAO](./tf_dao.md) section.
+- Transferring TFTs on different chains from the [TF Token Bridge](./tf_token_bridge.md) section.
+- Transferring TFTs on the TFChain from the [TF Token Transfer](./tf_token_transfer.md) section.
+- getting miniting reports from the [TF Minting Reports](./tf_minting_reports.md) section.
+
+ ![](../img/sidebar_4.png)
+
+***
+## Table of Content
+
+- [Your Profile](./your_profile.md)
+- [TF DAO](./tf_dao.md)
+- [TF Token Bridge](./tf_token_bridge.md)
+- [TF Token Transfer](./tf_token_transfer.md)
+- [TF Minting Reports](./tf_minting_reports.md)
\ No newline at end of file
diff --git a/collections/documentation/dashboard/tfchain/your_profile.md b/collections/documentation/dashboard/tfchain/your_profile.md
new file mode 100644
index 0000000..091cffe
--- /dev/null
+++ b/collections/documentation/dashboard/tfchain/your_profile.md
@@ -0,0 +1,13 @@
+Twin Management
+
+The TF Twin management feature of the ThreeFold Dashboard enables users to create, manage, and monitor their individual digital entities known as **Twins**. A Twin can represent a virtual machine (VM) or a container running on the ThreeFold Grid. With the Twin management, users can easily deploy and scale their workloads, allocate resources, and configure networking and storage settings for their Twins.
+
+![](../img/twin.png)
+
+The twin details consists of three main items.
+
+- `ID` Your unique identifier for your twin on the ThreeFold chain.
+- `Address` Your public address on the ThreeFold chain.
+- `Relay` A relay is a component that facilitates the reliable and secure transfer of messages between different entities within the ThreeFold ecosystem.
+
+To create a twin check the [Wallet Connector](../wallet_connector.md) Section.
\ No newline at end of file
diff --git a/collections/documentation/dashboard/tfgrid/grid_status.md b/collections/documentation/dashboard/tfgrid/grid_status.md
new file mode 100644
index 0000000..5969a99
--- /dev/null
+++ b/collections/documentation/dashboard/tfgrid/grid_status.md
@@ -0,0 +1,5 @@
+# Grid Status
+
+Check the status and health of ThreeFold services From [Grid Status](https://status.grid.tf/status/threefold)
+
+![](../img/grid_health.png)
\ No newline at end of file
diff --git a/collections/documentation/dashboard/tfgrid/node_monitoring.md b/collections/documentation/dashboard/tfgrid/node_monitoring.md
new file mode 100644
index 0000000..a7c8d51
--- /dev/null
+++ b/collections/documentation/dashboard/tfgrid/node_monitoring.md
@@ -0,0 +1,5 @@
+# Node Monitoring
+
+Monitor and check the metrics and status of Zero-OS nodes from [Node Monitoring](https://metrics.grid.tf/d/rYdddlPWkfqwf/zos-host-metrics?orgId=2&refresh=30s)
+
+![](../img/Monitoring.png)
\ No newline at end of file
diff --git a/collections/documentation/dashboard/tfgrid/node_statistics.md b/collections/documentation/dashboard/tfgrid/node_statistics.md
new file mode 100644
index 0000000..0980105
--- /dev/null
+++ b/collections/documentation/dashboard/tfgrid/node_statistics.md
@@ -0,0 +1,18 @@
+# Statistics
+
+Statistics allows you to see the distribution of 3Nodes all over the world with information on how many nodes are available and in which country.
+
+![](../img/statistics.png)
+
+Here you can see generic overview about:
+
+- Number of Nodes
+- Number of Dedicated Nodes
+- Number of Farms
+- Number of Countries
+- The capacity CRU, SRU, HRU, MRU
+- Number of GPUs
+- Number of Gateways
+- Number of Twins
+- The number of public IPs available
+- Number of Contracts
\ No newline at end of file
diff --git a/collections/documentation/dashboard/tfgrid/tfgrid.md b/collections/documentation/dashboard/tfgrid/tfgrid.md
new file mode 100644
index 0000000..2b06d9b
--- /dev/null
+++ b/collections/documentation/dashboard/tfgrid/tfgrid.md
@@ -0,0 +1,17 @@
+# TFGrid
+
+Check and use all things related to the threefold grid. Including:
+
+- The status of ThreeFold services from the [Grid Status](./grid_status.md) website.
+- The statistics of all nodes that are available on the ThreeFold grid from [Node Statistics](./node_statistics.md).
+- The health and status of Zero-OS nodes that are available on the ThreeFold grid from [Node Monitoring](./node_monitoring.md).
+
+ ![](../img/sidebar_1.png)
+
+***
+
+## Table of Content
+
+- [Grid Status](./grid_status.md)
+- [Node Statistics](./node_statistics.md)
+- [Node Monitoring](./node_monitoring.md)
\ No newline at end of file
diff --git a/collections/documentation/dashboard/toc.md b/collections/documentation/dashboard/toc.md
new file mode 100644
index 0000000..5507842
--- /dev/null
+++ b/collections/documentation/dashboard/toc.md
@@ -0,0 +1,8 @@
+## Dashboard TOC
+
+- [Home](./home.md)
+- [Wallet Connector](./wallet_connector.md)
+- [CapRover](./caprover.md)
+- [Virtual Machine](./vm.md)
+- [Funkwhale](./funkwhale.md)
+- [Peertube](./peertube.md)
diff --git a/collections/documentation/dashboard/vm_presearch.md b/collections/documentation/dashboard/vm_presearch.md
new file mode 100644
index 0000000..f438507
--- /dev/null
+++ b/collections/documentation/dashboard/vm_presearch.md
@@ -0,0 +1,51 @@
+# Mount a Presearch node on TFGrid3 using VM
+
+The fastest way to mount a Presearch node on TFGrid3 is inside a VM.
+
+Steps :
+- Set up a VM, see [here](./vm.md). It is recommended to reserve a fix IP. You can also try out the planetary network (so reserve a VM without public IP), as long as the node you select is connected to the internet through an IPv4 address that isn't used yet for a Presearch node, you don't explicitly need to reserve a public IPv4 address. However, the planetary network is still in beta phase and might generate performance issues.
+- 1 CPU is enough for a PRE node. As we still need to install Docker on the VM before deploying a PRE node, please choose 8192 memory size.
+- Once your VM is set up, SSH into our machine.
+
+![](./img/vm_list.png)
+
+For the VM having an IP address you can enter the terminal command
+```
+ssh root@185.206.122.162
+```
+
+If you didn't reserve a public IPv4 address, you can ssh into your machine using the IPv6 address. Peopl doing this, should however first set up their identity in the Planetary Network with Yggdrasil. See [here](yggdrasil_client) to know how to do this.
+
+```
+ssh root@300:aa1b:e91b:720f:ae5f:8991:6df8:1ec9
+```
+
+Now you have an empty VM, also Docker still needs to be installed.
+
+Execute the following commands inside your VM :
+
+```
+apt update ;
+apt install sudo ;
+sudo apt install apt-transport-https ca-certificates curl software-properties-common ;
+curl -fsSL https://download.docker.com/linux/ubuntu/gpg | sudo apt-key add - ;
+sudo add-apt-repository "deb [arch=amd64] https://download.docker.com/linux/ubuntu focal stable" ;
+```
+Finally, install Docker. The Ubuntu machine does not come with `systemd`. The following does the trick :
+
+```
+apt-get install docker-ce docker-ce-cli containerd.io
+apt-cache madison docker-ce
+
+dockerd &
+```
+
+Once Docker is set up, you can launch the PRE node instructions on your VM:
+
+```
+docker stop presearch-node ; docker rm presearch-node ; docker stop presearch-auto-updater ; docker rm presearch-auto-updater ; docker run -d --name presearch-auto-updater --restart=unless-stopped -v /var/run/docker.sock:/var/run/docker.sock presearch/auto-updater --cleanup --interval 900 presearch-auto-updater presearch-node ; docker pull presearch/node ; docker run -dt --name presearch-node --restart=unless-stopped -v presearch-node-storage:/app/node -e REGISTRATION_CODE=$YOUR_REGISTRATION_CODE_HERE presearch/node ; docker logs -f presearch-node
+```
+
+And you're done !
+
+![ ](./img/weblet_vm_presearch_result.jpg)
\ No newline at end of file
diff --git a/collections/documentation/dashboard/wallet_connector.md b/collections/documentation/dashboard/wallet_connector.md
new file mode 100644
index 0000000..842339a
--- /dev/null
+++ b/collections/documentation/dashboard/wallet_connector.md
@@ -0,0 +1,42 @@
+ Wallet Connector
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Supported Networks](#supported-networks)
+- [Process](#process)
+
+***
+
+## Introduction
+
+To interact with TFChain, users need to set a wallet connector.
+
+## Supported Networks
+
+Currently, we're supporting four different networks:
+
+- Dev net, for development purposes
+ - [https://dashboard.dev.grid.tf](https://dashboard.dev.grid.tf)
+- QA net, for internal testing and verifications
+ - [https://dashboard.qa.grid.tf](https://dashboard.qa.grid.tf)
+- Test net, for testing purposes
+ - [https://dashboard.test.grid.tf](https://dashboard.test.grid.tf)
+- Main net, for production-ready purposes
+ - [https://dashboard.grid.tf](https://dashboard.grid.tf)
+
+![ ](./img/profile_manager1.png)
+
+## Process
+
+Start entering the following information required to create your new profile.
+
+![ ](./img/profile_manager2.png)
+
+- `Mnemonics` are the secret words of your Polkadot account. Click on the **Create Account** button to generate yours.
+- `Password` is used to access your account
+- `Confirm Password`
+
+After you finish typing your credentials, click on **Connect**. Once your profile gets activated, you should find your **Twin ID** and **Address** generated under your **_Mnemonics_** for verification. Also, your **Account Balance** will be available at the top right corner under your profile name.
+
+![ ](./img/profile_manager3.png)
diff --git a/collections/documentation/developers/developers.md b/collections/documentation/developers/developers.md
new file mode 100644
index 0000000..3877320
--- /dev/null
+++ b/collections/documentation/developers/developers.md
@@ -0,0 +1,90 @@
+# ThreeFold Developers
+
+This section covers all practical tutorials on how to develop and build on the ThreeFold Grid.
+
+For complementary information on the technology developed by ThreeFold, refer to the [Technology](../../knowledge_base/technology/technology_toc.md) section.
+
+ Table of Contents
+
+- [Javascript Client](./javascript/grid3_javascript_readme.md)
+ - [Installation](./javascript/grid3_javascript_installation.md)
+ - [Loading Client](./javascript/grid3_javascript_loadclient.md)
+ - [Deploy a VM](./javascript/grid3_javascript_vm.md)
+ - [Capacity Planning](./javascript/grid3_javascript_capacity_planning.md)
+ - [Deploy Multiple VMs](./javascript/grid3_javascript_vms.md)
+ - [Deploy CapRover](./javascript/grid3_javascript_caprover.md)
+ - [Gateways](./javascript/grid3_javascript_vm_gateways.md)
+ - [Deploy a Kubernetes Cluster](./javascript/grid3_javascript_kubernetes.md)
+ - [Deploy a ZDB](./javascript/grid3_javascript_zdb.md)
+ - [Deploy ZDBs for QSFS](./javascript/grid3_javascript_qsfs_zdbs.md)
+ - [QSFS](./javascript/grid3_javascript_qsfs.md)
+ - [Key Value Store](./javascript/grid3_javascript_kvstore.md)
+ - [VM with Wireguard and Gateway](./javascript/grid3_wireguard_gateway.md)
+ - [GPU Support](./javascript/grid3_javascript_gpu_support.md)
+- [Go Client](./go/grid3_go_readme.md)
+ - [Installation](./go/grid3_go_installation.md)
+ - [Loading Client](./go/grid3_go_load_client.md)
+ - [Deploy a VM](./go/grid3_go_vm.md)
+ - [Deploy Multiple VMs](./go/grid3_go_vms.md)
+ - [Deploy Gateways](./go/grid3_go_gateways.md)
+ - [Deploy Kubernetes](./go/grid3_go_kubernetes.md)
+ - [Deploy a QSFS](./go/grid3_go_qsfs.md)
+ - [GPU Support](./go/grid3_go_gpu.md)
+- [TFCMD](./tfcmd/tfcmd.md)
+ - [Getting Started](./tfcmd/tfcmd_basics.md)
+ - [Deploy a VM](./tfcmd/tfcmd_vm.md)
+ - [Deploy Kubernetes](./tfcmd/tfcmd_kubernetes.md)
+ - [Deploy ZDB](./tfcmd/tfcmd_zdbs.md)
+ - [Gateway FQDN](./tfcmd/tfcmd_gateway_fqdn.md)
+ - [Gateway Name](./tfcmd/tfcmd_gateway_name.md)
+ - [Contracts](./tfcmd/tfcmd_contracts.md)
+- [TFROBOT](./tfrobot/tfrobot.md)
+ - [Installation](./tfrobot/tfrobot_installation.md)
+ - [Configuration File](./tfrobot/tfrobot_config.md)
+ - [Deployment](./tfrobot/tfrobot_deploy.md)
+ - [Commands and Flags](./tfrobot/tfrobot_commands_flags.md)
+ - [Supported Configurations](./tfrobot/tfrobot_configurations.md)
+- [ThreeFold Chain](./tfchain/tfchain.md)
+ - [Introduction](./tfchain/introduction.md)
+ - [Farming Policies](./tfchain/farming_policies.md)
+ - [External Service Contract](./tfchain/tfchain_external_service_contract.md)
+ - [Solution Provider](./tfchain/tfchain_solution_provider.md)
+- [Grid Proxy](./proxy/proxy_readme.md)
+ - [Introducing Grid Proxy](./proxy/proxy.md)
+ - [Setup](./proxy/setup.md)
+ - [DB Testing](./proxy/db_testing.md)
+ - [Commands](./proxy/commands.md)
+ - [Contributions](./proxy/contributions.md)
+ - [Explorer](./proxy/explorer.md)
+ - [Database](./proxy/database.md)
+ - [Production](./proxy/production.md)
+ - [Release](./proxy/release.md)
+- [Flist](./flist/flist.md)
+ - [ThreeFold Hub Intro](./flist/flist_hub/zos_hub.md)
+ - [Generate an API Token](./flist/flist_hub/api_token.md)
+ - [Convert Docker Image Into Flist](./flist/flist_hub/convert_docker_image.md)
+ - [Supported Flists](./flist/grid3_supported_flists.md)
+ - [Flist Case Studies](./flist/flist_case_studies/flist_case_studies.md)
+ - [Case Study: Debian 12](./flist/flist_case_studies/flist_debian_case_study.md)
+ - [Case Study: Nextcloud AIO](./flist/flist_case_studies/flist_nextcloud_case_study.md)
+- [Internals](./internals/internals.md)
+ - [Reliable Message Bus (RMB)](./internals/rmb/rmb_toc.md)
+ - [Introduction to RMB](./internals/rmb/rmb_intro.md)
+ - [RMB Specs](./internals/rmb/rmb_specs.md)
+ - [RMB Peer](./internals/rmb/uml/peer.md)
+ - [RMB Relay](./internals/rmb/uml/relay.md)
+ - [ZOS](./internals/zos/index.md)
+ - [Manual](./internals/zos/manual/manual.md)
+ - [Workload Types](./internals/zos/manual/workload_types.md)
+ - [Internal Modules](./internals/zos/internals/internals.md)
+ - [Capacity](./internals/zos/internals/capacity.md)
+ - [Performance Monitor Package](./internals/zos/performance/performance.md)
+ - [Public IPs Validation Task](./internals/zos/performance/publicips.md)
+ - [CPUBenchmark](./internals/zos/performance/cpubench.md)
+ - [IPerf](./internals/zos/performance/iperf.md)
+ - [Health Check](./internals/zos/performance/healthcheck.md)
+ - [API](./internals/zos/manual/api.md)
+- [Grid Deployment](./grid_deployment/grid_deployment.md)
+ - [TFGrid Stacks](./grid_deployment/tfgrid_stacks.md)
+ - [Full VM Grid Deployment](./grid_deployment/grid_deployment_full_vm.md)
+ - [Grid Snapshots](./grid_deployment/snapshots.md)
\ No newline at end of file
diff --git a/collections/documentation/developers/flist/flist.md b/collections/documentation/developers/flist/flist.md
new file mode 100644
index 0000000..6c69e05
--- /dev/null
+++ b/collections/documentation/developers/flist/flist.md
@@ -0,0 +1,11 @@
+ Flist
+
+ Table of Contents
+
+- [Zero-OS Hub](./flist_hub/zos_hub.md)
+- [Generate an API Token](./flist_hub/api_token.md)
+- [Convert Docker Image Into Flist](./flist_hub/convert_docker_image.md)
+- [Supported Flists](./grid3_supported_flists.md)
+- [Flist Case Studies](./flist_case_studies/flist_case_studies.md)
+ - [Case Study: Debian 12](./flist_case_studies/flist_debian_case_study.md)
+ - [Case Study: Nextcloud AIO](./flist_case_studies/flist_nextcloud_case_study.md)
\ No newline at end of file
diff --git a/collections/documentation/developers/flist/flist_case_studies/flist_case_studies.md b/collections/documentation/developers/flist/flist_case_studies/flist_case_studies.md
new file mode 100644
index 0000000..b258836
--- /dev/null
+++ b/collections/documentation/developers/flist/flist_case_studies/flist_case_studies.md
@@ -0,0 +1,6 @@
+ Flist Case Studies
+
+ Table of Contents
+
+- [Case Study: Debian 12](./flist_debian_case_study.md)
+- [Case Study: Nextcloud AIO](./flist_nextcloud_case_study.md)
\ No newline at end of file
diff --git a/collections/documentation/developers/flist/flist_case_studies/flist_debian_case_study.md b/collections/documentation/developers/flist/flist_case_studies/flist_debian_case_study.md
new file mode 100644
index 0000000..3777433
--- /dev/null
+++ b/collections/documentation/developers/flist/flist_case_studies/flist_debian_case_study.md
@@ -0,0 +1,300 @@
+ Flist Case Study: Debian 12
+
+ Table of Contents
+
+- [Introduction](#introduction)
+ - [You Said Flist?](#you-said-flist)
+ - [Case Study Objective](#case-study-objective)
+ - [The Overall Process](#the-overall-process)
+- [Docker Image Creation](#docker-image-creation)
+ - [Dockerfile](#dockerfile)
+ - [Docker Image Script](#docker-image-script)
+ - [zinit Folder](#zinit-folder)
+ - [README.md File](#readmemd-file)
+ - [Putting it All Together](#putting-it-all-together)
+- [Docker Publishing Steps](#docker-publishing-steps)
+ - [Create Account and Access Token](#create-account-and-access-token)
+ - [Build and Push the Docker Image](#build-and-push-the-docker-image)
+- [Convert the Docker Image to an Flist](#convert-the-docker-image-to-an-flist)
+- [Deploy the Flist on the TF Playground](#deploy-the-flist-on-the-tf-playground)
+- [Conclusion](#conclusion)
+
+***
+
+## Introduction
+
+For this tutorial, we will present a case study demonstrating how easy it is to create a new flist on the ThreeFold ecosystem. We will be creating a Debian Flist and we will deploy a micro VM on the ThreeFold Playground and access our Debian deployment.
+
+To do all this, we will need to create a Docker Hub account, create a Dockerfile, a docker image and a docker container, then convert the docker image to a Zero-OS flist. After all this, we will be deploying our Debian workload on the ThreeFold Playground. You'll see, it's pretty straightforward and fun to do.
+
+
+
+### You Said Flist?
+
+First, let's recall what an flist actually is and does. In short, an flist is a very effective way to deal with software data and the end result is fast deployment and high reliability.
+
+In a flist, we separate the metadata from the data. The metadata is a description of what files are in that particular image. It's the data providing information about the app/software. Thanks to flist, the 3Node doesn't need to install a complete software program in order to run properly. Only the necessary files are installed. Zero-OS can read the metadata of a container and only download and execute the necessary binaries and applications to run the workload, when it is necessary.
+
+Sounds great? It really is great, and very effective!
+
+One amazing thing about the flist technology is that it is possible to convert any Docker image into an flist, thanks to the [ThreeFold Docker Hub Converter tool](https://hub.grid.tf/docker-convert). If this sounds complicated, fear not. It is very easy and we will show you how to proceed in this case study.
+
+
+
+### Case Study Objective
+
+The goal of this case study is to give you enough information and tools so that you can yourself build your own flist projects and deploy on the ThreeFold Grid.
+
+This case study is not meant to show you all the detailed steps on creating an flist from scratch. We will instead start with some files templates available on the ThreeFold repository [tf-images](https://github.com/threefoldtech/tf-images). This is one of the many advantages of working with open-source projects: we can easily get inspiration from the already available codes of the many ThreeFold repositories and work our way up from there.
+
+
+
+### The Overall Process
+
+To give you a bird's view of the whole project, here are the main steps:
+
+* Create the Docker image
+* Push the Docker image to the Docker Hub
+* Convert the Docker image to a Zero-OS flist
+* Deploy a micro VM with the flist on the ThreeFold Playground
+
+
+
+## Docker Image Creation
+
+As we've said previously, we will not explore all the details of creating an flist from scratch. This would be done in a subsequent guide. For now, we want to take existing codes and work our way from there. This is not only quicker, but it is a good way to get to know the ThreeFold's ecosystem and repositories.
+
+We will be using the code available on the [ThreeFold Tech's Github page](https://github.com/threefoldtech). In our case, we want to explore the repository [tf-images](https://github.com/threefoldtech/tf-images).
+
+If you go on the subsection [tfgrid3](https://github.com/threefoldtech/tf-images/tree/development/tfgrid3), you can see many different flists available. In our case, we want to deploy the Debian Linux distribution. It is thus logic to try and find similar Linux distributions to take inspiration from.
+
+For this case study, we draw inspiration from the [Ubuntu 22.04](https://github.com/threefoldtech/tf-images/tree/development/tfgrid3/ubuntu22.04) directory.
+
+If we look at the Ubuntu 22.04 directory tree, this is what we get:
+
+```
+.
+├── Dockerfile
+├── README.md
+├── start.sh
+└── zinit
+ ├── ssh-init.yaml
+ └── sshd.yaml
+```
+
+We will now explore each of those files to get a good look at the whole repository and try to understand how it all works together.
+
+### Dockerfile
+
+We recall that to make a Docker image, you need to create a Dockerfile. As per [Docker's documentation](https://docs.docker.com/engine/reference/builder/), a Dockerfile is "a Dockerfile is a text document that contains all the commands a user could call on the command line to assemble an image".
+
+The Ubuntu 22.04 Dockerfile is as follows:
+
+File: `Dockerfile`
+
+```Dockerfile
+FROM ubuntu:22.04
+
+RUN apt update && \
+ apt -y install wget openssh-server
+
+RUN wget -O /sbin/zinit https://github.com/threefoldtech/zinit/releases/download/v0.2.5/zinit && \
+ chmod +x /sbin/zinit
+
+COPY zinit /etc/zinit
+COPY start.sh /start.sh
+
+RUN chmod +x /sbin/zinit && chmod +x /start.sh
+ENTRYPOINT ["zinit", "init"]
+```
+
+We can see from the first line that the Dockerfile will look for the docker image `ubuntu:22.04`. In our case, we want to get the Debian 12 docker image. This information is available on the Docker Hub (see [Debian Docker Hub](https://hub.docker.com/_/debian)).
+
+We will thus need to change the line `FROM ubuntu:22.04` to the line `FROM debian:12`. It isn't more complicated than that!
+
+We now have the following Dockerfile fore the Debian docker image:
+
+File: `Dockerfile`
+
+```Dockerfile
+FROM debian:12
+
+RUN apt update && \
+ apt -y install wget openssh-server
+
+RUN wget -O /sbin/zinit https://github.com/threefoldtech/zinit/releases/download/v0.2.5/zinit && \
+ chmod +x /sbin/zinit
+
+COPY zinit /etc/zinit
+COPY start.sh /start.sh
+
+RUN chmod +x /sbin/zinit && chmod +x /start.sh
+ENTRYPOINT ["zinit", "init"]
+```
+
+There is nothing more needed here. Pretty fun to start from some existing open-source code, right?
+
+### Docker Image Script
+
+The other important file we will be looking at is the `start.sh` file. This is the basic script that will be used to properly set the docker image. Thankfully, there is nothing more to change in this file, we can leave it as is. As we will see later, this file will be executed by zinit when the container starts.
+
+File: `start.sh`
+
+```.sh
+#!/bin/bash
+
+mkdir -p /var/run/sshd
+mkdir -p /root/.ssh
+touch /root/.ssh/authorized_keys
+
+chmod 700 /root/.ssh
+chmod 600 /root/.ssh/authorized_keys
+
+echo "$SSH_KEY" >> /root/.ssh/authorized_keys
+```
+
+### zinit Folder
+
+Next, we want to take a look at the zinit folder.
+
+But first, what is zinit? In a nutshell, zinit is a process manager (pid 1) that knows how to launch, monitor and sort dependencies. It thus executes targets in the proper order. For more information on zinit, check the [zinit repository](https://github.com/threefoldtech/zinit).
+
+When we start the Docker container, the files in the folder zinit will be executed.
+
+If we take a look at the file `ssh-init.yaml`, we find the following:
+
+```.yaml
+exec: bash /start.sh
+log: stdout
+oneshot: true
+````
+
+We can see that the first line calls the [bash](https://www.gnu.org/software/bash/) Unix shell and that it will run the file `start.sh` we've seen earlier.
+
+In this zinit service file, we define a service named `ssh-init.yaml`, where we tell zinit which commands to execute (here `bash /start.sh`), where to log (here in `stdout`) and where `oneshot` is set to `true` (meaning that it should only be executed once).
+
+If we take a look at the file `sshd.yaml`, we find the following:
+
+```.yaml
+exec: bash -c "/usr/sbin/sshd -D"
+after:
+ - ssh-init
+```
+
+Here another service `sshd.yaml` runs after the `ssh-init.yaml` process.
+
+### README.md File
+
+As every good programmer knows, a good code is nothing without some good documentation to help others understand what's going on! This is where the `README.md` file comes into play.
+
+In this file, we can explain what our code is doing and offer steps to properly configure the whole deployment. For the users that will want to deploy the flist on the ThreeFold Playground, they would need the FLIst URL and the basic steps to deploy a Micro VM on the TFGrid. We will thus add this information in the README.md file. This information can be seen in the [section below](#deploy-the-flist-on-the-tf-playground). To read the complete README.md file, go to [this link](https://github.com/threefoldtech/tf-images/tree/development/tfgrid3/debian).
+
+### Putting it All Together
+
+We've now went through all the files available in the Ubuntu 22.04 directory on the tf-images repository. To build your own image, you would simply need to put all those files in a local folder on your computer and follow the steps presented at the next section, [Docker Publishing Steps](#docker-publishing-steps).
+
+To have a look at the final result of the changes we bring to the Ubuntu 22.04 version, have a look at this [Debian directory](https://github.com/threefoldtech/tf-images/tree/development/tfgrid3/debian) on the ThreeFold's tf-images repository.
+
+
+
+## Docker Publishing Steps
+
+### Create Account and Access Token
+
+To be able to push Docker images to the Docker Hub, you obviously need to create a Docker Hub account! This is very easy and please note that there are so many amazing documentation on Docker online. If you're lost, make the most of your favorite search engine and find a way out of the blue.
+
+Here are the steps to create an account and an access token.
+
+* Go to the [Docker Hub](https://hub.docker.com/)
+* Click `Register` and follow the steps given by Docker
+* On the top right corner, click on your account name and select `Account Settings`
+* On the left menu, click on `Security`
+* Click on `New Access Token`
+* Choose an Access Token description that you will easily identify then click `Generate`
+ * Make sure to set the permissions `Read, Write, Delete`
+* Follow the steps given to properly connect your local computer to the Docker Hub
+ * Run `docker login -u `
+ * Set the password
+
+You now have access to the Docker Hub from your local computer. We will then proceed to push the Docker image we've created.
+
+### Build and Push the Docker Image
+
+* Make sure the Docker Daemon is running
+* Build the docker container
+ * Template:
+ * ```
+ docker build -t / .
+ ```
+ * Example:
+ * ```
+ docker build -t username/debian12 .
+ ```
+* Push the docker container to the [Docker Hub](https://hub.docker.com/)
+ * Template:
+ * ```
+ docker push /
+ ```
+ * Example:
+ * ```
+ docker push username/debian12
+ ```
+* You should now see your docker image on the [Docker Hub](https://hub.docker.com/) when you go into the menu option `My Profile`.
+ * Note that you can access this link quickly with the following template:
+ * ```
+ https://hub.docker.com/u/
+ ```
+
+
+
+## Convert the Docker Image to an Flist
+
+We will now convert the Docker image into a Zero-OS flist. This part is so easy you will almost be wondering why you never heard about flist before!
+
+* Go to the [ThreeFold Hub](https://hub.grid.tf/).
+* Sign in with the ThreeFold Connect app.
+* Go to the [Docker Hub Converter](https://hub.grid.tf/docker-convert) section.
+* Next to `Docker Image Name`, add the docker image repository and name, see the example below:
+ * Template:
+ * `/docker_image_name:tagname`
+ * Example:
+ * `username/debian12:latest`
+* Click `Convert the docker image`.
+* Once the conversion is done, the flist is available as a public link on the ThreeFold Hub.
+* To get the flist URL, go to the [TF Hub main page](https://hub.grid.tf/), scroll down to your 3Bot ID and click on it.
+* Under `Name`, you will see all your available flists.
+* Right-click on the flist you want and select `Copy Clean Link`. This URL will be used when deploying on the ThreeFold Playground. We show below the template and an example of what the flist URL looks like.
+ * Template:
+ * ```
+ https://hub.grid.tf/<3BOT_name.3bot>/--.flist
+ ```
+ * Example:
+ * ```
+ https://hub.grid.tf/idrnd.3bot/username-debian12-latest.flist
+ ```
+
+
+
+## Deploy the Flist on the TF Playground
+
+* Go to the [ThreeFold Playground](https://play.grid.tf).
+* Set your profile manager.
+* Go to the [Micro VM](https://play.grid.tf/#/vm) page.
+* Choose your parameters (name, VM specs, etc.).
+* Under `flist`, paste the Debian flist from the TF Hub you copied previously.
+* Make sure the entrypoint is as follows:
+ * ```
+ /sbin/zinit init
+ ```
+* Choose a 3Node to deploy on
+* Click `Deploy`
+
+That's it! You can now SSH into your Debian deployment and change the world one line of code at a time!
+
+*
+
+## Conclusion
+
+In this case study, we've seen the overall process of creating a new flist to deploy a Debian workload on a Micro VM on the ThreeFold Playground.
+
+If you have any questions or feedback, please let us know by either writing a post on the [ThreeFold Forum](https://forum.threefold.io/), or by chatting with us on the [TF Grid Tester Community](https://t.me/threefoldtesting) Telegram channel.
diff --git a/collections/documentation/developers/flist/flist_case_studies/flist_nextcloud_case_study.md b/collections/documentation/developers/flist/flist_case_studies/flist_nextcloud_case_study.md
new file mode 100644
index 0000000..4193317
--- /dev/null
+++ b/collections/documentation/developers/flist/flist_case_studies/flist_nextcloud_case_study.md
@@ -0,0 +1,858 @@
+ Flist Case Study: Nextcloud All-in-One
+
+ Table of Contents
+
+- [Introduction](#introduction)
+ - [Flist: What is It?](#flist-what-is-it)
+ - [Case Study Objective](#case-study-objective)
+ - [The Overall Process](#the-overall-process)
+- [Docker Image Creation](#docker-image-creation)
+ - [Nextcloud Flist Directory Tree](#nextcloud-flist-directory-tree)
+ - [Caddyfile](#caddyfile)
+ - [Dockerfile](#dockerfile)
+ - [README.md File](#readmemd-file)
+ - [scripts Folder](#scripts-folder)
+ - [caddy.sh](#caddysh)
+ - [sshd\_init.sh](#sshd_initsh)
+ - [ufw\_init.sh](#ufw_initsh)
+ - [nextcloud.sh](#nextcloudsh)
+ - [nextcloud\_conf.sh](#nextcloud_confsh)
+ - [zinit Folder](#zinit-folder)
+ - [ssh-init.yaml and sshd.yaml](#ssh-inityaml-and-sshdyaml)
+ - [ufw-init.yaml and ufw.yaml](#ufw-inityaml-and-ufwyaml)
+ - [caddy.yaml](#caddyyaml)
+ - [dockerd.yaml](#dockerdyaml)
+ - [nextcloud.yaml](#nextcloudyaml)
+ - [nextcloud-conf.yaml](#nextcloud-confyaml)
+ - [Putting it All Together](#putting-it-all-together)
+- [Docker Publishing Steps](#docker-publishing-steps)
+ - [Create Account and Access Token](#create-account-and-access-token)
+ - [Build and Push the Docker Image](#build-and-push-the-docker-image)
+- [Convert the Docker Image to an Flist](#convert-the-docker-image-to-an-flist)
+- [Deploy Nextcloud AIO on the TFGrid with Terraform](#deploy-nextcloud-aio-on-the-tfgrid-with-terraform)
+ - [Create the Terraform Files](#create-the-terraform-files)
+ - [Deploy Nextcloud with Terraform](#deploy-nextcloud-with-terraform)
+ - [Nextcloud Setup](#nextcloud-setup)
+- [Conclusion](#conclusion)
+
+***
+
+# Introduction
+
+In this case study, we explain how to create a new flist on the ThreeFold ecosystem. We will show the process of creating a Nextcloud All-in-One flist and we will deploy a micro VM on the ThreeFold Playground to access our Nextcloud instance. As a reference, the official Nextcloud flist is available [here](https://hub.grid.tf/tf-official-apps/threefoldtech-nextcloudaio-latest.flist.md).
+
+To achieve all this, we will need to create a Docker Hub account, create a Dockerfile and its associated files, a docker image and a docker container, then convert the docker image to a Zero-OS flist. After all this, we will be deploying our Nextcloud instance on the ThreeFold Playground.
+
+As a general advice, before creating an flist for a ThreeFold deployment, you should make sure that you are able to deploy your workload properly by using a micro VM or a full VM on the TFGrid. Once you know all the steps to deploy your workload, and after some thorough tests, you can take what you've learned and incorporate all this into an flist.
+
+## Flist: What is It?
+
+Before we go any further, let us recall what is an flist. In short, an flist is a technology for storing and efficiently sharing sets of files. While it has many great features, it's purpose in this case is simply to deliver the image contents to Zero-OS for execution as a micro VM. It thus acts as a bundle of files like a normal archive.
+
+One convenient thing about the flist technology is that it is possible to convert any Docker image into an flist, thanks to the [ThreeFold Docker Hub Converter tool](https://hub.grid.tf/docker-convert). It is very easy to do and we will show you how to proceed in this case study. For a quick guide on converting Docker images into flists, read [this section](../flist_hub/convert_docker_image.md) of the ThreeFold Manual.
+
+## Case Study Objective
+
+The goal of this case study is to give you enough information and tools so that you can build your own flist projects and deploy on the ThreeFold Grid.
+
+We will explore the different files needed to create the flist and explain the overall process. Instead of starting from scratch, we will analyze the Nextcloud flist directory in the [tf-images](https://github.com/threefoldtech/tf-images/tree/development/tfgrid3/nextcloud) ThreeFold Tech repository. As the project is already done, it will be easier to get an overview of the process and the different components so you can learn to create your own.
+
+## The Overall Process
+
+To give you a bird's-eye view of the whole project, here are the main steps:
+
+* Create the Docker image
+* Push the Docker image to the Docker Hub
+* Convert the Docker image to a Zero-OS flist
+* Deploy a micro VM with the flist on the ThreeFold Playground with Terraform
+
+One important thing to have in mind is that, when we create an flist, what we are doing is basically automating the required steps to deploy a given workload on the TFGrid. Usually, these steps would be done manually and step-by-step by an individual deploying on a micro or a full VM.
+
+Once we've successfully created an flist, we thus have a very quick way to deploy a specific workload while always obtaining the same result. This is why it is highly recommended to test a given deployment on a full or micro VM before building an flist.
+
+For example, in the case of building a Nextcloud All-in-One flist, the prerequisites would be to successfully deploy a Nextcloud AIO instance on a full VM by executing each step sequentially. This specific example is documented in the Terraform section [Nextcloud All-in-One Guide](../../../system_administrators/terraform/advanced/terraform_nextcloud_aio.md) of the System Administrators book.
+
+In our case, the flist we will be using has some specific configurations depending on the way we deploy Nextcloud (e.g. using or not the gateway and a custom domain). The Terraform **main.tf** we will be sharing later on will thus take all this into account for a smooth deployment.
+
+# Docker Image Creation
+
+As we've said previously, we will explore the different components of the existing Nextcloud flist directory. We thus want to check the existing files and try to understand as much as possible how the different components work together. This is also a very good introduction to the ThreeFold ecosystem.
+
+We will be using the files available on the [ThreeFold Tech Github page](https://github.com/threefoldtech). In our case, we want to explore the repository [tf-images](https://github.com/threefoldtech/tf-images).
+
+If you go in the subsection [tfgrid3](https://github.com/threefoldtech/tf-images/tree/development/tfgrid3), you can see many different flists available. In our case, we want to deploy the [Nextcloud All-in-One Flist](https://github.com/threefoldtech/tf-images/tree/development/tfgrid3/nextcloud).
+
+## Nextcloud Flist Directory Tree
+
+The Nextcloud flist directory tree is the following:
+
+```
+tree tf-images/tfgrid3/nextcloud
+.
+├── Caddyfile
+├── Dockerfile
+├── README.md
+├── scripts
+│ ├── caddy.sh
+│ ├── nextcloud_conf.sh
+│ ├── nextcloud.sh
+│ ├── sshd_init.sh
+│ └── ufw_init.sh
+└── zinit
+ ├── caddy.yaml
+ ├── dockerd.yaml
+ ├── nextcloud-conf.yaml
+ ├── nextcloud.yaml
+ ├── sshd.yaml
+ ├── ssh-init.yaml
+ ├── ufw-init.yaml
+ └── ufw.yaml
+```
+
+We can see that the directory is composed of a Caddyfile, a Dockerfile, a README.md and two directories, **scripts** and **zinit**. We will now explore each of those components to have a good grasp of the whole repository and to understand how it all works together.
+
+To get a big picture of this directory, we could say that the **README.md** file provides the necessary documentation for the users to understand the Nextcloud flist, how it is built and how it works, the **Caddyfile** provides the necessary requirements to run the reverse proxy, the **Dockerfile** specifies how the Docker image is built, installing things such as [openssh](https://www.openssh.com/) and the [ufw firewall](https://wiki.ubuntu.com/UncomplicatedFirewall) for secure remote connection, while the two folders, **scripts** and **zinit**, could be said to work hand-in-hand.
+
+Each `.yaml` file is a *unit file* for zinit. That means it specifies a single service for zinit to start. We'll learn more about these files later, but for now we can just note that each script file (ending with `.sh`) has an associated zinit file to make sure that the script is run. There are also some other files for running programs aside from our scripts.
+
+## Caddyfile
+
+For our Nextcloud deployment, we are using Caddy as a reverse proxy. A reverse proxy is an application that sits in front of back-end applications and forwards client requests to those applications.
+
+Since Nextcloud AIO actually includes two web applications, both Nextcloud itself and the AIO management interface, we use the reverse proxy to serve them both on a single domain. It also allows us to make some changes on the fly to the content of the AIO site to considerably enhance the user experience. Finally, we also use Caddy to provide SSL termination if the user reserves a public IP and no gateway, since otherwise SSL termination is provided by the gateway.
+
+File: `Caddyfile`
+```
+{
+ order replace after encode
+ servers {
+ trusted_proxies static 100.64.0.0/10 10.0.0.0/8
+ }
+}
+
+
+{$DOMAIN}:{$PORT} {
+ handle_path /aio* {
+ replace {
+ href="/ href="/aio/
+ src="/ src="/aio/
+ action=" action="/aio
+ url(' url('/aio
+ `value="" placeholder="nextcloud.yourdomain.com"` `value="{$DOMAIN}"`
+ `"Submit domain"` `"Submit domain" id="domain-submit"`
+ {$REPLACEMENTS}
+ {$BODY}
+ }
+
+ reverse_proxy localhost:8000 {
+ header_down Location "^/(.*)$" "/aio/$1"
+ header_down Refresh "^/(.*)$" "/aio/$1"
+ }
+
+ }
+
+ redir /api/auth/getlogin /aio{uri}
+
+ reverse_proxy localhost:11000
+
+ handle_errors {
+ @502-aio expression {err.status_code} == 502 && path('/aio*')
+ handle @502-aio {
+ header Content-Type text/html
+ respond <
+ Nextcloud
+ Your Nextcloud management interface isn't ready. If you just deployed this instance, please wait a minute and refresh the page.
+
+ HTML 200
+ }
+
+ @502 expression {err.status_code} == 502
+ handle @502 {
+ redir /* /aio
+ }
+ }
+}
+```
+
+We can see in the first section (`trusted_proxies static`) that we set a range of IP addresses as trusted proxy addresses. These include the possible source addresses for gateway traffic, which we mark as trusted for compatibility with some Nextcloud features.
+
+After the global config at the top, the line `{$DOMAIN}:{$PORT}` defines the port that Caddy will listen to and the domain that we are using for our site. This is important, because in the case that port `443` is specified, Caddy will handle SSL certificates automatically.
+
+The following blocks define behavior for different URL paths that users might try to access.
+
+To begin, we have `/aio*`. This is how we place the AIO management app in a "subfolder" of our main domain. To accomplish that we need a few rules that rewrite the contents of the returned pages to correct the links. We also add some text replacements here to accomplish the enhancements mentioned earlier, like automatically filling the domain entry field.
+
+With the `reverse_proxy` line, we specify that requests to all URLs starting with `/aio` should be sent to the web server running on port `8000` of `localhost`. That's the port where the AIO server is listening, as we'll see below. There's also a couple of header rewrite rules here that correct the links for any redirects the AIO site makes.
+
+The `redir` line is needed to support a feature where users open the AIO interface from within Nextcloud. This redirects the original request to the correct equivalent within the `/aio` "subfolder".
+
+Then there's a second `reverse_proxy` line, which is the catch-all for any traffic that didn't get intercepted earlier. This handles the actual Nextcloud app and sends the traffic to its separate server running on port `11000`.
+
+The section starting with `handle_errors` ensures that the user will receive an understandable error message when trying to access the Nextcloud deployment before it has fully started up.
+
+## Dockerfile
+
+We recall that to make a Docker image, you need to create a Dockerfile. As per the [Docker documentation](https://docs.docker.com/engine/reference/builder/), a Dockerfile is "a text document that contains all the commands a user could call on the command line to assemble an image".
+
+File: `Dockerfile`
+
+```Dockerfile
+FROM ubuntu:22.04
+
+RUN apt update && \
+ apt -y install wget openssh-server curl sudo ufw inotify-tools iproute2
+
+RUN wget -O /sbin/zinit https://github.com/threefoldtech/zinit/releases/download/v0.2.5/zinit && \
+ chmod +x /sbin/zinit
+
+RUN wget -O /sbin/caddy 'https://caddyserver.com/api/download?os=linux&arch=amd64&p=github.com%2Fcaddyserver%2Freplace-response&idempotency=43631173212363' && \
+ chmod +x /sbin/caddy
+
+RUN curl -fsSL https://get.docker.com -o /usr/local/bin/install-docker.sh && \
+ chmod +x /usr/local/bin/install-docker.sh
+
+RUN sh /usr/local/bin/install-docker.sh
+
+COPY ./Caddyfile /etc/caddy/
+COPY ./scripts/ /scripts/
+COPY ./zinit/ /etc/zinit/
+RUN chmod +x /scripts/*.sh
+
+ENTRYPOINT ["/sbin/zinit", "init"]
+```
+
+We can see from the first line that this Dockerfile uses a base image of Ubuntu Linux version 22.04.
+
+With the first **RUN** command, we refresh the package lists, and then install **openssh**, **ufw** and other dependencies for our Nextcloud uses. Note that we also install **curl** so that we can quickly install **Docker**.
+
+With the second **RUN** command, we install **zinit** and we give it execution permission with the command `chmod +x`. More will be said about zinit in a section below.
+
+With the third **RUN** command, we install **caddy** and we give it execution permission with the command `chmod +x`. Caddy is an extensible, cross-platform, open-source web server written in Go. For more information on Caddy, check the [Caddy website](https://caddyserver.com/).
+
+With fourth **RUN** command, we download and give proper permissions to the script `install-docker.sh`. On a terminal, the common line to install Docker would be `curl -fsSL https://get.docker.com | sudo sh`. To understand really what's going here, we can simply go to the link provided at the line [https://get.docker.com](https://get.docker.com) for more information.
+
+The fifth **RUN** command runs the `install-docker.sh` script to properly install Docker within the image.
+
+Once those commands are run, we proceed to copy into our Docker image the necessary folders `scripts` and `zinit` as well as the Caddyfile. Once this is done, we give execution permissions to all scripts in the scripts folder using `chmod +x`.
+
+Finally, we set an entrypoint in our Dockerfile. As per the [Docker documentation](https://docs.docker.com/engine/reference/builder/), an entrypoint "allows you to configure a container that will run as an executable". Since we are using zinit, we set the entrypoint `/sbin/zinit`.
+
+## README.md File
+
+The **README.md** file has the main goal of explaining clearly to the user the functioning of the Nextcloud directory and its associated flist. In this file, we can explain what our code is doing and offer steps to properly configure the whole deployment.
+
+We also give the necessary steps to create the Docker image and convert it into an flist starting directly with the Nextcloud directory. This can be useful for users that want to create their own flist, instead of using the [official ThreeFold Nextcloud flist](https://hub.grid.tf/tf-official-apps/threefoldtech-nextcloudaio-latest.flist.md).
+
+To read the complete README.md file, go to [this link](https://github.com/threefoldtech/tf-images/blob/development/tfgrid3/nextcloud/README.md).
+
+## scripts Folder
+
+The **scripts** folder contains without surprise the scripts necessary to run the Nextcloud instance.
+
+In the Nextcloud Flist case, there are five scripts:
+
+* **caddy.sh**
+* **nextcloud.sh**
+* **nextcloud_conf.sh**
+* **sshd_init.sh**
+* **ufw_init.sh**
+
+Let's take a look at each of them.
+
+### caddy.sh
+
+File: `caddy.sh`
+
+```bash
+#!/bin/bash
+export DOMAIN=$NEXTCLOUD_DOMAIN
+
+if $IPV4 && ! $GATEWAY; then
+ export PORT=443
+else
+ export PORT=80
+fi
+
+if $IPV4; then
+ export BODY="\`\`"
+
+else
+ export BODY="\`\`"
+
+ export REPLACEMENTS=' `name="talk"` `name="talk" disabled`
+ `needs ports 3478/TCP and 3478/UDP open/forwarded in your firewall/router` `running the Talk container requires a public IP and this VM does not have one. It is still possible to use Talk in a limited capacity. Please consult the documentation for details`'
+fi
+
+caddy run --config /etc/caddy/Caddyfile
+```
+
+The script **caddy.sh** sets the proper port depending on the network configuration (e.g. IPv4 or Gateway) in the first if/else section. In the second if/else section, the script also makes sure that the proper domain is given to Nextcloud All-in-One. This quickens the installation process as the user doesn't have to set the domain in Nextcloud AIO after deployment. We also disable a feature that's not relevant if the user didn't reserve an IPv4 address and we insert a note about that.
+
+### sshd_init.sh
+
+File: `sshd_init.sh`
+
+```bash
+#!/bin/bash
+
+mkdir -p ~/.ssh
+mkdir -p /var/run/sshd
+chmod 600 ~/.ssh
+chmod 600 /etc/ssh/*
+echo $SSH_KEY >> ~/.ssh/authorized_keys
+```
+
+This file starts with a shebang (`#!`) that instructs the operating system to execute the following lines using the [Bash shell](https://www.gnu.org/software/bash/). In essence, it lets us write `./sshd_init.sh` with the same outcome as `bash ./sshd_init.sh`, assuming the file is executable.
+
+The goal of this script is to add the public key within the VM in order for the user to get a secure and remote connection to the VM. The two lines starting with `mkdir` create the necessary folders. The lines starting with `chmod` give the owner the permission to write and read the content within the folders. Finally, the line `echo` will write the public SSH key in a file within the VM. In the case that the flist is used as a weblet, the SSH key is set in the Playground profile manager and passed as an environment variable when we deploy the solution.
+
+### ufw_init.sh
+
+File: `ufw_init.sh`
+
+```bash
+#!/bin/bash
+
+ufw default deny incoming
+ufw default allow outgoing
+ufw allow ssh
+ufw allow http
+ufw allow https
+ufw allow 8443
+ufw allow 3478
+ufw limit ssh
+```
+
+The goal of the `ufw_init.sh` script is to set the correct firewall parameters to make sure that our deployment is secure while also providing the necessary access for the Nextcloud users.
+
+The first two lines starting with `ufw default` are self-explanatory. We want to restrain incoming traffic while making sure that outgoing traffic has no restraints.
+
+The lines starting with `ufw allow` open the ports necessary for our Nextcloud instance. We note that **ssh** is port 22, **http** is port 80 and **https** is port 443. This means, for example, that the line `ufw allow 22` is equivalent to the line `ufw allow ssh`.
+
+Port 8443 can be used to access the AIO interface, as an alternative to using the `/aio` "subfolder" on deployments with a public IPv4 address. Finally, the port 3478 is used for Nextcloud Talk.
+
+The line `ufw limit ssh` will provide additional security by denying connection from IP addresses that attempt to initiate 6 or more connections within a 30-second period.
+
+### nextcloud.sh
+
+File: `nextcloud.sh`
+
+```bash
+#!/bin/bash
+
+export COMPOSE_HTTP_TIMEOUT=800
+while ! docker info > /dev/null 2>&1; do
+ echo docker not ready
+ sleep 2
+done
+
+docker run \
+--init \
+--sig-proxy=false \
+--name nextcloud-aio-mastercontainer \
+--restart always \
+--publish 8000:8000 \
+--publish 8080:8080 \
+--env APACHE_PORT=11000 \
+--env APACHE_IP_BINDING=0.0.0.0 \
+--env SKIP_DOMAIN_VALIDATION=true \
+--volume nextcloud_aio_mastercontainer:/mnt/docker-aio-config \
+--volume /var/run/docker.sock:/var/run/docker.sock:ro \
+nextcloud/all-in-one:latest
+```
+
+The **nextcloud.sh** script is where the real action starts. This is where we run the Nextcloud All-in-One docker image.
+
+Before discussing the main part of this script, we note that the `while` loop is used to ensure that the `docker run` command starts only after the Docker daemon has properly started.
+
+The code section starting with `docker run` is taken from the [Nextcloud All-in-One repository on Github](https://github.com/nextcloud/all-in-one) with some slight modifications. The last line indicates that the Docker image being pulled will always be the latest version of Nextcloud All-in-One.
+
+We note here that Nextcloud AIO is published on the port 8000 and 8080. We also note that we set restart to **always**. This is very important as it will make sure that the Nextcloud instance is restarted if the Docker daemon reboots. We take the opportunity to note that the way zinit configures micro VMs, the Docker daemon restarts automatically after a reboot. Thus, this latter fact combined with the line `--restart always` ensures that the user that the Nextcloud instance will restart after a VM reboot.
+
+We also set **11000** as the Apache port with an IP binding of **0.0.0.0**. For our deployment, we want to skip the domain validation, thus it is set to **true**.
+
+Considering the line `--sig-proxy=false`, when this command is run interactively, it prevents the user from accidentally killing the spawned AIO container. While it is not of great importance in our case, it means that zinit will not kill the container if the service is stopped.
+
+For more information on this, we invite the readers to consult the [Nextcloud documentation](https://github.com/nextcloud/all-in-one#how-to-use-this).
+
+### nextcloud_conf.sh
+
+File: `nextcloud_conf.sh`
+
+```bash
+#!/bin/bash
+
+# Wait for the nextcloud container to become healthy. Note that we can set the
+# richtext config parameters even before the app is installed
+
+nc_ready () {
+ until [[ "`docker inspect -f {{.State.Health.Status}} nextcloud-aio-nextcloud 2> /dev/null`" == "healthy" ]]; do
+ sleep 1;
+ done;
+}
+
+# When a gateway is used, AIO sets the WOPI allow list to only include the
+# gateway IP. Since requests don't originate from the gateway IP, they are
+# blocked by default. Here we add the public IP of the VM, or of the router
+# upstream of the node
+# See: github.com/nextcloud/security-advisories/security/advisories/GHSA-24x8-h6m2-9jf2
+
+if $IPV4; then
+ interface=$(ip route show default | cut -d " " -f 5)
+ ipv4_address=$(ip a show $interface | grep -Po 'inet \K[\d.]+')
+fi
+
+if $GATEWAY; then
+ nc_ready
+ wopi_list=$(docker exec --user www-data nextcloud-aio-nextcloud php occ config:app:get richdocuments wopi_allowlist)
+
+ if $IPV4; then
+ ip=$ipv4_address
+ else
+ ip=$(curl -fs https://ipinfo.io/ip)
+ fi
+
+ if [[ $ip ]] && ! echo $wopi_list | grep -q $ip; then
+ docker exec --user www-data nextcloud-aio-nextcloud php occ config:app:set richdocuments wopi_allowlist --value=$ip
+ fi
+fi
+
+
+# If the VM has a gateway and a public IPv4, then AIO will set the STUN/TURN
+# servers to the gateway domain which does not point to the public IP, so we
+# use the IP instead. In this case, we must wait for the Talk app to be
+# installed before changing the settings. With inotifywait, we don't need
+# a busy loop that could run indefinitely
+
+apps_dir=/mnt/data/docker/volumes/nextcloud_aio_nextcloud/_data/custom_apps/
+
+if $GATEWAY && $IPV4; then
+ if [[ ! -d ${apps_dir}spreed ]]; then
+ inotifywait -qq -e create --include spreed $apps_dir
+ fi
+ nc_ready
+
+ turn_list=$(docker exec --user www-data nextcloud-aio-nextcloud php occ talk:turn:list)
+ turn_secret=$(echo "$turn_list" | grep secret | cut -d " " -f 4)
+ turn_server=$(echo "$turn_list" | grep server | cut -d " " -f 4)
+
+ if ! echo $turn_server | grep -q $ipv4_address; then
+ docker exec --user www-data nextcloud-aio-nextcloud php occ talk:turn:delete turn $turn_server udp,tcp
+ docker exec --user www-data nextcloud-aio-nextcloud php occ talk:turn:add turn $ipv4_address:3478 udp,tcp --secret=$turn_secret
+ fi
+
+ stun_list=$(docker exec --user www-data nextcloud-aio-nextcloud php occ talk:stun:list)
+ stun_server=$(echo $stun_list | cut -d " " -f 2)
+
+ if ! echo $stun_server | grep -q $ipv4_address; then
+ docker exec --user www-data nextcloud-aio-nextcloud php occ talk:stun:add $ipv4_address:3478
+ docker exec --user www-data nextcloud-aio-nextcloud php occ talk:stun:delete $stun_server
+ fi
+fi
+```
+
+The script **nextcloud_conf.sh** ensures that the network settings are properly configured. In the first section, we use a function called **nc_ready ()**. This function will makes sure that the rest of the script only starts when the Nextcloud container is healthy.
+
+We note that the comments present in this script explain very well what is happening. In short, we want to set the Nextcloud instance according to the user's choice of network. For example, the user can decide to deploy using a ThreeFold gateway or a standard IPv4 connection. If the VM has a gateway and a public IPv4, then Nextcloud All-in-One will set the STUN/TURN servers to the gateway domain which does not point to the public IP, so we use the IP instead.
+
+## zinit Folder
+
+Next, we want to take a look at the zinit folder.
+
+But first, what is zinit? In a nutshell, zinit is a process manager (pid 1) that knows how to launch, monitor and sort dependencies. It thus executes targets in the proper order. For more information on zinit, check the [zinit repository](https://github.com/threefoldtech/zinit).
+
+When we start the Docker container, zinit will parse each unit file in the `/etc/zinit` folder and execute the contained command according to the specified parameters.
+
+In the Nextcloud Flist case, there are eight **.yaml** files:
+
+* **caddy.yaml**
+* **dockerd.yaml**
+* **nextcloud-conf.yaml**
+* **nextcloud.yaml**
+* **ssh-init.yaml**
+* **sshd.yaml**
+* **ufw-init.yaml**
+* **ufw.yaml**
+
+
+### ssh-init.yaml and sshd.yaml
+
+We start by taking a look at the **ssh-init.yaml** and **sshd.yaml** files.
+
+File: `ssh-init.yaml`
+
+```yaml
+exec: /scripts/sshd_init.sh
+oneshot: true
+```
+
+In this zinit service file, we define a service named `ssh-init.yaml`, where we tell zinit to execute the following command: `exec: /scripts/sshd_init.sh`. This unit file thus runs the script `sshd_init.sh` we covered in a previous section.
+
+We also note that `oneshot` is set to `true` and this means that it should only be executed once. This directive is often used for setup scripts that only need to run once. When it is not specified, the default value of `false` means that zinit will continue to start up a service if it ever dies.
+
+Now, we take a look at the file `sshd.yaml`:
+
+File: `sshd.yaml`
+
+```yaml
+exec: bash -c "/usr/sbin/sshd -D"
+after:
+ - ssh-init
+```
+
+We can see that this file executes a line from the Bash shell. It is important to note that, with zinit and .yaml files, you can easily order the executions of the files with the `after` directive. In this case, it means that the service `sshd` will only run after `ssh-init`.
+
+### ufw-init.yaml and ufw.yaml
+
+Let's take a look at the files **ufw-init.yaml** and **ufw.yaml**.
+
+File: `ufw-init.yaml`
+
+```yaml
+exec: /scripts/ufw_init.sh
+oneshot: true
+```
+
+The file `ufw-init.yaml` is very similar to the previous file `ssh-init.yaml`.
+
+File: `ufw.yaml`
+
+```yaml
+exec: ufw --force enable
+oneshot: true
+after:
+ - ufw-init
+```
+
+We can see that the file `ufw.yaml` will only run once and only after the file `ufw-init.yaml` has been run. This is important since the file `ufw-init.yaml` executes the script `ufw_init.sh`. We recall this script allows different ports in the firewall. Once those ports are defined, we can then run the command `ufw --force enable`. This will start the ufw firewall.
+
+### caddy.yaml
+
+```yaml
+exec: /scripts/caddy.sh
+oneshot: true
+```
+
+This is also very similar to previous files and just runs the Caddy script as a oneshot.
+
+### dockerd.yaml
+
+We now take a look at the file **dockerd.yaml**.
+
+File: `dockerd.yaml`
+
+```yaml
+exec: /usr/bin/dockerd --data-root /mnt/data/docker
+```
+
+This file will run the [dockerd daemon](https://docs.docker.com/engine/reference/commandline/dockerd/) which is the persistent process that manages containers. We also note that it sets the data to be stored in the directory **/mnt/data/docker**, which is important because we will mount a virtual disk there that will provide better performance, especially for Docker's storage driver.
+
+### nextcloud.yaml
+
+File: `nextcloud.yaml`
+
+```yaml
+exec: /scripts/nextcloud.sh
+after:
+ - dockerd
+```
+
+The file `nextcloud.yaml` runs after dockerd.
+
+This file will execute the `nextcloud.sh` script we saw earlier. We recall that this script starts the Nextcloud All-in-One image.
+
+### nextcloud-conf.yaml
+
+File: `nextcloud-conf.yaml`
+
+```yaml
+exec: /scripts/nextcloud_conf.sh
+oneshot: true
+after:
+ - nextcloud
+```
+
+Finally, the file `nextcloud-conf.yaml` runs after `nextcloud.yaml`.
+
+This file will execute the `nextcloud-conf.sh` script we saw earlier. We recall that this script starts the Nextcloud All-in-One image. At this point, the deployment is complete.
+
+## Putting it All Together
+
+We've now gone through all the files in the Nextcloud flist directory. You should now have a proper understanding of the interplay between the zinit (.yaml) and the scripts (.sh) files as well as the basic steps to build a Dockerfile and to write clear documentation.
+
+To build your own Nextcloud docker image, you would simply need to clone this directory to your local computer and to follow the steps presented in the next section [Docker Publishing Steps](#docker-publishing-steps).
+
+To have a look at the complete directory, you can always refer to the [Nextcloud flist directory](https://github.com/threefoldtech/tf-images/tree/development/tfgrid3/nextcloud) on the ThreeFold tf-images repository.
+
+# Docker Publishing Steps
+
+In this section, we show the necessary steps to publish the Docker image to the Docker Hub.
+
+To do so, we need to create an account and an access token. Then we will build the Docker image and push it to the Docker Hub.
+
+## Create Account and Access Token
+
+To be able to push Docker images to the Docker Hub, you obviously need to create a Docker Hub account! This is very easy and note that there are many great tutorials online about Docker.
+
+Here are the steps to create an account and an access token:
+
+* Go to the [Docker Hub](https://hub.docker.com/)
+* Click `Register` and follow the steps given by Docker
+* On the top right corner, click on your account name and select `Account Settings`
+* On the left menu, click on `Security`
+* Click on `New Access Token`
+* Choose an Access Token description that you will easily identify then click `Generate`
+ * Make sure to set the permissions `Read, Write, Delete`
+* On your local computer, make sure that the Docker daemon is running
+* Write the following in the command line to connect to the Docker hub:
+ * Run `docker login -u `
+ * Set the password
+
+You now have access to the Docker Hub from your local computer. We will then proceed to push the Docker image to the Docker Hub.
+
+## Build and Push the Docker Image
+
+* Make sure the Docker Daemon is running
+* Build the docker container (note that, while the tag is optional, it can help to track different versions)
+ * Template:
+ * ```
+ docker build -t /: .
+ ```
+ * Example:
+ * ```
+ docker build -t dockerhubuser/nextcloudaio .
+ ```
+* Push the docker container to the [Docker Hub](https://hub.docker.com/)
+ * Template:
+ * ```
+ docker push /
+ ```
+ * Example:
+ * ```
+ docker push dockerhubuser/nextcloudaio
+ ```
+* You should now see your docker image on the [Docker Hub](https://hub.docker.com/) when you go into the menu option `My Profile`.
+ * Note that you can access this link quickly with the following template:
+ * ```
+ https://hub.docker.com/u/
+ ```
+
+# Convert the Docker Image to an Flist
+
+We will now convert the Docker image into a Zero-OS flist.
+
+* Go to the [ThreeFold Hub](https://hub.grid.tf/).
+* Sign in with the ThreeFold Connect app.
+* Go to the [Docker Hub Converter](https://hub.grid.tf/docker-convert) section.
+* Next to `Docker Image Name`, add the docker image repository and name, see the example below:
+ * Template:
+ * `/docker_image_name:tagname`
+ * Example:
+ * `dockerhubuser/nextcloudaio:latest`
+* Click `Convert the docker image`.
+* Once the conversion is done, the flist is available as a public link on the ThreeFold Hub.
+* To get the flist URL, go to the [TF Hub main page](https://hub.grid.tf/), scroll down to your 3Bot ID and click on it.
+* Under `Name`, you will see all your available flists.
+* Right-click on the flist you want and select `Copy Clean Link`. This URL will be used when deploying on the ThreeFold Playground. We show below the template and an example of what the flist URL looks like.
+ * Template:
+ * ```
+ https://hub.grid.tf/<3BOT_name.3bot>/--.flist
+ ```
+ * Example:
+ * ```
+ https://hub.grid.tf/tf-official-apps/threefoldtech-nextcloudaio-latest.flist
+ ```
+
+# Deploy Nextcloud AIO on the TFGrid with Terraform
+
+We now proceed to deploy a Nextcloud All-in-One instance by using the Nextcloud flist we've just created.
+
+To do so, we will deploy a micro VM with the Nextcloud flist on the TFGrid using Terraform.
+
+## Create the Terraform Files
+
+For this guide, we use two files to deploy with Terraform. The first file contains the environment variables and the second file contains the parameters to deploy our workloads.
+
+To facilitate the deployment, only the environment variables file needs to be adjusted. The **main.tf** file contains the environment variables (e.g. **var.size** for the disk size) and thus you do not need to change this file. Of course, you can adjust the deployment based on your preferences. That being said, it should be easy to deploy the Terraform deployment with the main.tf as is.
+
+For this example, we will be deployment with a ThreeFold gateway as well as a gateway domain.
+
+* Copy the following content and save the file under the name `credentials.auto.tfvars`:
+
+```
+mnemonics = "..."
+network = "main"
+SSH_KEY = "..."
+
+size = "50"
+cpu = "2"
+memory = "4096"
+
+gateway_id = "50"
+vm1_id = "5453"
+
+deployment_name = "nextcloudgateway"
+nextcloud_flist = "https://hub.grid.tf/tf-official-apps/threefoldtech-nextcloudaio-latest.flist"
+```
+
+Make sure to add your own seed phrase and SSH public key. Simply replace the three dots by the content. Note that you can deploy on a different node than node 5453 for the **vm1** node. If you want to deploy on another node than node 5453 for the **gateway** node, make sure that you choose a gateway node. To find a gateway node, go on the [ThreeFold Dashboard](https://dashboard.grid.tf/) Nodes section of the Explorer and select **Gateways (Only)**.
+
+Obviously, you can decide to increase or modify the quantity in the variables `size`, `cpu` and `memory`.
+
+Note that in our case, we set the flist to be the official Nextcloud flist. Simply replace the URL with your newly created Nextcloud flist to test it!
+
+* Copy the following content and save the file under the name `main.tf`:
+
+```
+variable "mnemonics" {
+ type = string
+ default = "your mnemonics"
+}
+
+variable "network" {
+ type = string
+ default = "main"
+}
+
+variable "SSH_KEY" {
+ type = string
+ default = "your SSH pub key"
+}
+
+variable "deployment_name" {
+ type = string
+}
+
+variable "size" {
+ type = string
+}
+
+variable "cpu" {
+ type = string
+}
+
+variable "memory" {
+ type = string
+}
+
+variable "nextcloud_flist" {
+ type = string
+}
+
+variable "gateway_id" {
+ type = string
+}
+
+variable "vm1_id" {
+ type = string
+}
+
+
+terraform {
+ required_providers {
+ grid = {
+ source = "threefoldtech/grid"
+ }
+ }
+}
+
+provider "grid" {
+ mnemonics = var.mnemonics
+ network = var.network
+}
+
+data "grid_gateway_domain" "domain" {
+ node = var.gateway_id
+ name = var.deployment_name
+}
+
+resource "grid_network" "net" {
+ nodes = [var.gateway_id, var.vm1_id]
+ ip_range = "10.1.0.0/16"
+ name = "network"
+ description = "My network"
+ add_wg_access = true
+}
+
+resource "grid_deployment" "d1" {
+ node = var.vm1_id
+ network_name = grid_network.net.name
+
+ disks {
+ name = "data"
+ size = var.size
+ }
+
+ vms {
+ name = "vm1"
+ flist = var.nextcloud_flist
+ cpu = var.cpu
+ memory = var.memory
+ rootfs_size = 15000
+ entrypoint = "/sbin/zinit init"
+ env_vars = {
+ SSH_KEY = var.SSH_KEY
+ GATEWAY = "true"
+ IPV4 = "false"
+ NEXTCLOUD_DOMAIN = data.grid_gateway_domain.domain.fqdn
+ }
+ mounts {
+ disk_name = "data"
+ mount_point = "/mnt/data"
+ }
+ }
+}
+
+resource "grid_name_proxy" "p1" {
+ node = var.gateway_id
+ name = data.grid_gateway_domain.domain.name
+ backends = [format("http://%s:80", grid_deployment.d1.vms[0].ip)]
+ network = grid_network.net.name
+ tls_passthrough = false
+}
+
+output "wg_config" {
+ value = grid_network.net.access_wg_config
+}
+
+output "vm1_ip" {
+ value = grid_deployment.d1.vms[0].ip
+}
+output "vm1_ygg_ip" {
+ value = grid_deployment.d1.vms[0].ygg_ip
+}
+
+output "fqdn" {
+ value = data.grid_gateway_domain.domain.fqdn
+}
+```
+
+## Deploy Nextcloud with Terraform
+
+We now deploy Nextcloud with Terraform. Make sure that you are in the correct folder containing the main and variables files.
+
+* Initialize Terraform:
+ * ```
+ terraform init
+ ```
+
+* Apply Terraform to deploy Nextcloud:
+ * ```
+ terraform apply
+ ```
+
+Note that, at any moment, if you want to see the information on your Terraform deployment, write the following:
+ * ```
+ terraform show
+ ```
+
+## Nextcloud Setup
+
+Once you've deployed Nextcloud, you can access the Nextcloud setup page by pasting the URL displayed on the line `fqdn = "..."` of the Terraform output.
+
+# Conclusion
+
+In this case study, we've seen the overall process of creating a new flist to deploy a Nextcloud instance on a Micro VM on the TFGrid with Terraform.
+
+If you have any questions or feedback, please let us know by either writing a post on the [ThreeFold Forum](https://forum.threefold.io/), or by chatting with us on the [TF Grid Tester Community](https://t.me/threefoldtesting) Telegram channel.
\ No newline at end of file
diff --git a/collections/documentation/developers/flist/flist_case_studies/img/nextcloud_logo.jpeg b/collections/documentation/developers/flist/flist_case_studies/img/nextcloud_logo.jpeg
new file mode 100644
index 0000000..2d85227
Binary files /dev/null and b/collections/documentation/developers/flist/flist_case_studies/img/nextcloud_logo.jpeg differ
diff --git a/collections/documentation/developers/flist/flist_hub/api_token.md b/collections/documentation/developers/flist/flist_hub/api_token.md
new file mode 100644
index 0000000..5ca968f
--- /dev/null
+++ b/collections/documentation/developers/flist/flist_hub/api_token.md
@@ -0,0 +1,33 @@
+ TF Hub API Token
+
+ Table of Contents
+
+- [Generate an API Token](#generate-an-api-token)
+- [Verify the Token Validity](#verify-the-token-validity)
+
+***
+
+## Generate an API Token
+
+To generate an API Token on the TF Hub, follow those steps:
+
+* Go to the [ThreeFold Hub](https://hub.grid.tf/)
+* Open the top right drop-down menu
+* Click on `Generate API Token`
+* Take note of the token and keep it somewhere safe
+
+## Verify the Token Validity
+
+To make sure the generated token is valid, in the terminal write the following with your own API Token:
+
+```bash
+curl -H "Authorization: bearer " https://hub.grid.tf/api/flist/me
+```
+
+You should see the following line with your own 3BotID
+
+```bash
+{"status": "success", "payload": {"username": "<3BotID>.3bot"}}
+```
+
+You can then use this API Token in the terminal to [get and update information through the API](./zos_hub.md#get-and-update-information-through-the-api).
\ No newline at end of file
diff --git a/collections/documentation/developers/flist/flist_hub/convert_docker_image.md b/collections/documentation/developers/flist/flist_hub/convert_docker_image.md
new file mode 100644
index 0000000..e6fba85
--- /dev/null
+++ b/collections/documentation/developers/flist/flist_hub/convert_docker_image.md
@@ -0,0 +1,45 @@
+ Convert Docker Image to Flist
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Upload the Image](#upload-the-image)
+- [Flist on the Hub](#flist-on-the-hub)
+
+***
+
+## Introduction
+
+We show the steps to convert a docker image to an Flist.
+
+## Upload the Image
+
+1. Upload the Docker image to Docker Hub with the following command:
+
+```bash
+docker push
+```
+
+2. Navigate to the docker converter link: https://hub.grid.tf/docker-convert
+ ![ ](./img/docker_convert.png)
+
+3. Copy the name of the uploaded Docker image to the Docker Image Name field.
+
+4. Then press the convert button.
+
+When the image is ready, some information will be displayed.
+
+![ ](./img/flist_ready.png)
+
+## Flist on the Hub
+
+To navigate to the created flist, you can search with the newly created file name in the search tab.
+
+![ ](./img/search.png)
+
+You can also navigate to your repository in the contributors section from the Zero-Os Hub and navigate to the newly created flist.
+
+Then press the preview button to display the flist's url and some other data.
+
+![ ](./img/preview.png)
+
diff --git a/collections/documentation/developers/flist/flist_hub/img/docker_convert.png b/collections/documentation/developers/flist/flist_hub/img/docker_convert.png
new file mode 100644
index 0000000..1be37a0
Binary files /dev/null and b/collections/documentation/developers/flist/flist_hub/img/docker_convert.png differ
diff --git a/collections/documentation/developers/flist/flist_hub/img/flist_ready.png b/collections/documentation/developers/flist/flist_hub/img/flist_ready.png
new file mode 100644
index 0000000..c8913d5
Binary files /dev/null and b/collections/documentation/developers/flist/flist_hub/img/flist_ready.png differ
diff --git a/collections/documentation/developers/flist/flist_hub/img/hub_flist.png b/collections/documentation/developers/flist/flist_hub/img/hub_flist.png
new file mode 100644
index 0000000..3e5331a
Binary files /dev/null and b/collections/documentation/developers/flist/flist_hub/img/hub_flist.png differ
diff --git a/collections/documentation/developers/flist/flist_hub/img/preview.png b/collections/documentation/developers/flist/flist_hub/img/preview.png
new file mode 100644
index 0000000..bd555ca
Binary files /dev/null and b/collections/documentation/developers/flist/flist_hub/img/preview.png differ
diff --git a/collections/documentation/developers/flist/flist_hub/img/search.png b/collections/documentation/developers/flist/flist_hub/img/search.png
new file mode 100644
index 0000000..a5128cd
Binary files /dev/null and b/collections/documentation/developers/flist/flist_hub/img/search.png differ
diff --git a/collections/documentation/developers/flist/flist_hub/zos_hub.md b/collections/documentation/developers/flist/flist_hub/zos_hub.md
new file mode 100644
index 0000000..95bdbc8
--- /dev/null
+++ b/collections/documentation/developers/flist/flist_hub/zos_hub.md
@@ -0,0 +1,142 @@
+ Zero-OS Hub
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Upload Your Files](#upload-your-files)
+- [Merge Multiple Flists](#merge-multiple-flists)
+- [Convert Docker Images and Tar Files](#convert-docker-images-and-tar-files)
+- [Upload Customize Flists](#upload-customize-flists)
+- [Upload Homemade Flists](#upload-homemade-flists)
+- [Upload your Existing Flist to Reduce Bandwidth](#upload-your-existing-flist-to-reduce-bandwidth)
+- [Authenticate via 3Bot](#authenticate-via-3bot)
+- [Get and Update Information Through the API](#get-and-update-information-through-the-api)
+ - [Public API Endpoints (No Authentication Required)](#public-api-endpoints-no-authentication-required)
+ - [Restricted API Endpoints (Authentication Required)](#restricted-api-endpoints-authentication-required)
+ - [API Request Templates and Examples](#api-request-templates-and-examples)
+
+***
+
+## Introduction
+
+The [ThreeFold Zero-OS Hub](https://hub.grid.tf/) allows you to do multiple things and acts as a public centralization of flists.
+
+The ZOS Hub is mainly there to gives an easy way to distribute flist files, which are databases of metadata that you can use in any Zero-OS container or virtual machine.
+
+## Upload Your Files
+In order to publish easily your files, you can upload a `.tar.gz` and the hub will convert it automatically to a flist
+and store the contents in the hub backend. After that you can use your flist directly on a container.
+
+## Merge Multiple Flists
+In order to reduce the maintenance of your images, products, etc. flist allows you to keep your
+different products and files separately and then merge them with another flist to make it usable without
+keeping the system up-to-date.
+
+Example: there is an official `ubuntu 16.04` flist image, you can make a flist which contains your application files
+and then merge your flist with ubuntu, so the resulting flist is your product on the last version of ubunbu.
+You don't need to take care about the base system yourself, just merge it with the one provided.
+
+## Convert Docker Images and Tar Files
+
+The ZOS Hub allows you to convert Docker Hub images and Tar files into flists thanks to the Docker Hub Converter.
+
+You can convert a docker image (eg: `busybox`, `ubuntu`, `fedora`, `couchdb`, ...) to an flist directly from the backend, this allows you to use your existing docker image in our infrastructure out-of-the-box. Go to the [Docker Hub Converter](https://hub.grid.tf/docker-convert) to use this feature. For more information on the process, read the section [Convert Docker Image to flist](./convert_docker_image.md) of the TF Manual.
+
+You can also easily convert a Tar file into an flist via the [Upload section](https://hub.grid.tf/upload) of the ZOS Hub.
+
+## Upload Customize Flists
+
+The ZOS Hub also allows you to customize an flist via the [Customization section](https://hub.grid.tf/merge) of the ZOS Hub. Note that this is currently in beta.
+
+## Upload Homemade Flists
+
+The ZOS Hub allows you to upload flist that you've made yourself via the section [Upload a homemade flist](https://hub.grid.tf/upload-flist).
+
+## Upload your Existing Flist to Reduce Bandwidth
+In addition with the hub-client (a side product) you can upload efficiently contents of file
+to make the backend up-to-date and upload a self-made flist. This allows you to do all the jobs yourself
+and gives you the full control of the chain. The only restriction is that the contents of the files you host
+on the flist needs to exists on the backend, otherwise your flist will be rejected.
+
+## Authenticate via 3Bot
+All the operations on the ZOS Hub needs to be done via a `3Bot` (default) authentication. Only downloading a flist can be done anonymously. To authenticate request via the API, you need to generate an API Token as shown in the section [ZOS Hub API Token](./api_token.md).
+
+## Get and Update Information Through the API
+The hub host a basic REST API which can gives you some informations about flists, renaming them, remove them, etc.
+
+To use authenticated endpoints, you need to provide a itsyou.online valid `jwt` via `Authorization: bearer ` header.
+This `jwt` can contains special `memberof` to allows you cross-repository actions.
+
+If your `jwt` contains memberof, you can choose which user you want to use by specifying cookie `active-user`.
+See example below.
+
+
+### Public API Endpoints (No Authentication Required)
+- `/api/flist` (**GET**)
+ - Returns a json array with all repository/flists found
+- `/api/repositories` (**GET**)
+ - Returns a json array with all repositories found
+- `/api/fileslist` (**GET**)
+ - Returns a json array with all repositories and files found
+- `/api/flist/` (**GET**)
+ - Returns a json array of each flist found inside specified repository.
+ - Each entry contains `filename`, `size`, `updated` date and `type` (regular or symlink), optionally `target` if it's a symbolic link.
+- `/api/flist//` (**GET**)
+ - Returns json object with flist dumps (full file list)
+
+### Restricted API Endpoints (Authentication Required)
+- `/api/flist/me` (**GET**)
+ - Returns json object with some basic information about yourself (authenticated user)
+- `/api/flist/me/` (**GET**, **DELETE**)
+ - **GET**: same as `/api/flist//`
+ - **DELETE**: remove that specific flist
+- `/api/flist/me//link/` (**GET**)
+ - Create a symbolic link `linkname` pointing to `source`
+- `/api/flist/me//crosslink//` (**GET**)
+ - Create a cross-repository symbolic link `linkname` pointing to `repository/sourcename`
+- `/api/flist/me//rename/` (**GET**)
+ - Rename `source` to `destination`
+- `/api/flist/me/promote///` (**GET**)
+ - Copy cross-repository `sourcerepo/sourcefile` to your `[local-repository]/localname` file
+ - This is useful when you want to copy flist from one repository to another one, if your jwt allows it
+- `/api/flist/me/upload` (**POST**)
+ - **POST**: uploads a `.tar.gz` archive and convert it to an flist
+ - Your file needs to be passed via `file` form attribute
+- `/api/flist/me/upload-flist` (**POST**)
+ - **POST**: uploads a `.flist` file and store it
+ - Note: the flist is checked and full contents is verified to be found on the backend, if some chunks are missing, the file will be discarded.
+ - Your file needs to be passed via `file` form attribute
+- `/api/flist/me/merge/` (**POST**)
+ - **POST**: merge multiple flist together
+ - You need to passes a json array of flists (in form `repository/file`) as POST body
+- `/api/flist/me/docker` (**POST**)
+ - **POST**: converts a docker image to an flist
+ - You need to passes `image` form argument with docker-image name
+ - The resulting conversion will stay on your repository
+
+### API Request Templates and Examples
+
+The main template to request information from the API is the following:
+
+```bash
+curl -H "Authorization: bearer " https://hub.grid.tf/api/flist/me/ -X
+```
+
+For example, if we take the command `DELETE` of the previous section and we want to delete the flist `example-latest.flist` with the API Token `abc12`, we would write the following line:
+
+```bash
+curl -H "Authorization: bearer abc12" https://hub.grid.tf/api/flist/me/example-latest.flist -X DELETE
+```
+
+As another template example, if we wanted to rename the flist `current-name-latest.flist` to `new-name-latest.flist`, we would use the following template:
+
+```bash
+curl -H "Authorization: bearer " https://hub.grid.tf/api/flist/me//rename/ -X GET
+```
+
+To upload an flist to the ZOS Hub, you would use the following template:
+
+```bash
+curl -H "Authorization: bearer " -X POST -F file=@my-local-archive.tar.gz \
+ https://hub.grid.tf/api/flist/me/upload
+```
\ No newline at end of file
diff --git a/collections/documentation/developers/flist/grid3_supported_flists.md b/collections/documentation/developers/flist/grid3_supported_flists.md
new file mode 100644
index 0000000..537d0b3
--- /dev/null
+++ b/collections/documentation/developers/flist/grid3_supported_flists.md
@@ -0,0 +1,26 @@
+ Supported Flists
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Flists and Parameters](#flists-and-parameters)
+- [More Flists](#more-flists)
+
+***
+
+## Introduction
+
+We provide basic information on the currently supported Flists.
+
+## Flists and Parameters
+
+|flist|entrypoint|env vars|
+|:--:|:--:|--|
+|[Alpine](https://hub.grid.tf/tf-official-apps/threefoldtech-alpine-3.flist.md)|`/entrypoint.sh`|`SSH_KEY`|
+|[Ubuntu](https://hub.grid.tf/tf-official-apps/threefoldtech-ubuntu-22.04.flist.md)|`/init.sh`|`SSH_KEY`|
+|[CentOS](https://hub.grid.tf/tf-official-apps/threefoldtech-centos-8.flist.md)|`/entrypoint.sh`|`SSH_KEY`|
+|[K3s](https://hub.grid.tf/tf-official-apps/threefoldtech-k3s-latest.flist.md)|`/sbin/zinit init`|- `SSH_KEY` - `K3S_TOKEN` - `K3S_DATA_DIR` - `K3S_FLANNEL_IFACE` - `K3S_NODE_NAME` - `K3S_URL` `https://${masterIp}:6443`|
+
+## More Flists
+
+You can convert any docker image to an flist. Feel free to explore the different possibilities on the [ThreeFold Hub](https://hub.grid.tf/).
\ No newline at end of file
diff --git a/collections/documentation/developers/go/grid3_go_gateways.md b/collections/documentation/developers/go/grid3_go_gateways.md
new file mode 100644
index 0000000..66f5f68
--- /dev/null
+++ b/collections/documentation/developers/go/grid3_go_gateways.md
@@ -0,0 +1,104 @@
+ Deploying Gateways
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Gateway Name](#gateway-name)
+- [Example](#example)
+- [Gateway FQDN](#gateway-fqdn)
+- [Example](#example-1)
+
+***
+
+## Introduction
+
+After [deploying a VM](./grid3_go_vm.md) you can deploy Gateways to further expose your VM.
+
+## Gateway Name
+
+This generates a FQDN for your VM.
+
+## Example
+
+```go
+import (
+ "fmt"
+
+ "github.com/threefoldtech/tfgrid-sdk-go/grid-client/deployer"
+ "github.com/threefoldtech/tfgrid-sdk-go/grid-client/workloads"
+ "github.com/threefoldtech/tfgrid-sdk-go/grid-proxy/pkg/types"
+ "github.com/threefoldtech/zos/pkg/gridtypes/zos"
+)
+
+func main() {
+
+ // Create Threefold plugin client
+ tfPluginClient, err := deployer.NewTFPluginClient(mnemonics, "sr25519", network, "", "", true, false)
+
+ // Get a free node to deploy
+ domain := true
+ status := "up"
+ filter := types.NodeFilter{
+ Domain: &domain,
+ Status: &status,
+ }
+ nodeIDs, err := deployer.FilterNodes(tfPluginClient.GridProxyClient, filter)
+ nodeID := uint32(nodeIDs[0].NodeID)
+
+ // Create gateway to deploy
+ gateway := workloads.GatewayNameProxy{
+ NodeID: nodeID,
+ Name: "mydomain",
+ Backends: []zos.Backend{"http://[300:e9c4:9048:57cf:6d98:42c6:a7bf:2e3f]:8080"},
+ TLSPassthrough: true,
+ }
+ err = tfPluginClient.GatewayNameDeployer.Deploy(ctx, &gateway)
+
+ gatewayObj, err := tfPluginClient.State.LoadGatewayNameFromGrid(nodeID, gateway.Name, gateway.Name)
+ fmt.Println(gatewayObj.FQDN)
+}
+
+```
+
+This deploys a Gateway Name Proxy that forwards requests to your VM. You should see an output like this:
+
+```bash
+mydomain.gent01.dev.grid.tf
+```
+
+## Gateway FQDN
+
+In case you have a FQDN already pointing to the node, you can expose your VM using Gateway FQDN.
+
+## Example
+
+```go
+import (
+ "fmt"
+
+ "github.com/threefoldtech/tfgrid-sdk-go/grid-client/deployer"
+ "github.com/threefoldtech/tfgrid-sdk-go/grid-client/workloads"
+ "github.com/threefoldtech/zos/pkg/gridtypes/zos"
+)
+
+func main() {
+
+ // Create Threefold plugin client
+ tfPluginClient, err := deployer.NewTFPluginClient(mnemonics, "sr25519", network, "", "", "", 0, true)
+
+ // Create gateway to deploy
+ gateway := workloads.GatewayFQDNProxy{
+ NodeID: 14,
+ Name: "mydomain",
+ Backends: []zos.Backend{"http://[300:e9c4:9048:57cf:6d98:42c6:a7bf:2e3f]:8080"},
+ FQDN: "my.domain.com",
+ TLSPassthrough: true,
+ }
+ err = tfPluginClient.GatewayFQDNDeployer.Deploy(ctx, &gateway)
+
+ gatewayObj, err := tfPluginClient.State.LoadGatewayFQDNFromGrid(nodeID, gateway.Name, gateway.Name)
+}
+
+```
+
+This deploys a Gateway FQDN Proxy that forwards requests to from node 14 public IP to your VM.
diff --git a/collections/documentation/developers/go/grid3_go_gpu.md b/collections/documentation/developers/go/grid3_go_gpu.md
new file mode 100644
index 0000000..3ff0c94
--- /dev/null
+++ b/collections/documentation/developers/go/grid3_go_gpu.md
@@ -0,0 +1,6 @@
+ GPU and Go
+
+ Table of Contents
+
+- [GPU and Go Introduction](grid3_go_gpu_support.md)
+- [Deploy a VM with GPU](grid3_go_vm_with_gpu.md)
\ No newline at end of file
diff --git a/collections/documentation/developers/go/grid3_go_gpu_support.md b/collections/documentation/developers/go/grid3_go_gpu_support.md
new file mode 100644
index 0000000..9aadceb
--- /dev/null
+++ b/collections/documentation/developers/go/grid3_go_gpu_support.md
@@ -0,0 +1,116 @@
+ GPU Support
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Example](#example)
+- [More Information](#more-information)
+
+***
+
+## Introduction
+
+We present here an example on how to deploy using the Go client. This is part of our integration tests.
+
+
+
+## Example
+
+```go
+func TestVMWithGPUDeployment(t *testing.T) {
+ tfPluginClient, err := setup()
+ assert.NoError(t, err)
+
+ ctx, cancel := context.WithTimeout(context.Background(), 5*time.Minute)
+ defer cancel()
+
+ publicKey, privateKey, err := GenerateSSHKeyPair()
+ assert.NoError(t, err)
+
+ twinID := uint64(tfPluginClient.TwinID)
+ nodeFilter := types.NodeFilter{
+ Status: &statusUp,
+ FreeSRU: convertGBToBytes(20),
+ FreeMRU: convertGBToBytes(8),
+ RentedBy: &twinID,
+ HasGPU: &trueVal,
+ }
+
+ nodes, err := deployer.FilterNodes(ctx, tfPluginClient, nodeFilter)
+ if err != nil {
+ t.Skip("no available nodes found")
+ }
+ nodeID := uint32(nodes[0].NodeID)
+
+ nodeClient, err := tfPluginClient.NcPool.GetNodeClient(tfPluginClient.SubstrateConn, nodeID)
+ assert.NoError(t, err)
+
+ gpus, err := nodeClient.GPUs(ctx)
+ assert.NoError(t, err)
+
+ network := workloads.ZNet{
+ Name: "gpuNetwork",
+ Description: "network for testing gpu",
+ Nodes: []uint32{nodeID},
+ IPRange: gridtypes.NewIPNet(net.IPNet{
+ IP: net.IPv4(10, 20, 0, 0),
+ Mask: net.CIDRMask(16, 32),
+ }),
+ AddWGAccess: false,
+ }
+
+ disk := workloads.Disk{
+ Name: "gpuDisk",
+ SizeGB: 20,
+ }
+
+ vm := workloads.VM{
+ Name: "gpu",
+ Flist: "https://hub.grid.tf/tf-official-vms/ubuntu-22.04.flist",
+ CPU: 4,
+ Planetary: true,
+ Memory: 1024 * 8,
+ GPUs: ConvertGPUsToStr(gpus),
+ Entrypoint: "/init.sh",
+ EnvVars: map[string]string{
+ "SSH_KEY": publicKey,
+ },
+ Mounts: []workloads.Mount{
+ {DiskName: disk.Name, MountPoint: "/data"},
+ },
+ NetworkName: network.Name,
+ }
+
+ err = tfPluginClient.NetworkDeployer.Deploy(ctx, &network)
+ assert.NoError(t, err)
+
+ defer func() {
+ err = tfPluginClient.NetworkDeployer.Cancel(ctx, &network)
+ assert.NoError(t, err)
+ }()
+
+ dl := workloads.NewDeployment("gpu", nodeID, "", nil, network.Name, []workloads.Disk{disk}, nil, []workloads.VM{vm}, nil)
+ err = tfPluginClient.DeploymentDeployer.Deploy(ctx, &dl)
+ assert.NoError(t, err)
+
+ defer func() {
+ err = tfPluginClient.DeploymentDeployer.Cancel(ctx, &dl)
+ assert.NoError(t, err)
+ }()
+
+ vm, err = tfPluginClient.State.LoadVMFromGrid(nodeID, vm.Name, dl.Name)
+ assert.NoError(t, err)
+ assert.Equal(t, vm.GPUs, ConvertGPUsToStr(gpus))
+
+ time.Sleep(30 * time.Second)
+ output, err := RemoteRun("root", vm.YggIP, "lspci -v", privateKey)
+ assert.NoError(t, err)
+ assert.Contains(t, string(output), gpus[0].Vendor)
+}
+```
+
+
+
+## More Information
+
+For more information on this, you can check this [Client Pull Request](https://github.com/threefoldtech/tfgrid-sdk-go/pull/207/) on how to support the new calls to list GPUs and to deploy a machine with GPU.
\ No newline at end of file
diff --git a/collections/documentation/developers/go/grid3_go_installation.md b/collections/documentation/developers/go/grid3_go_installation.md
new file mode 100644
index 0000000..c07008c
--- /dev/null
+++ b/collections/documentation/developers/go/grid3_go_installation.md
@@ -0,0 +1,45 @@
+Go Client Installation
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Requirements](#requirements)
+- [Steps](#steps)
+- [References](#references)
+
+***
+
+## Introduction
+
+We present the general steps to install the ThreeFold Grid3 Go Client.
+
+## Requirements
+
+Make sure that you have at least Go 1.19 installed on your machine.
+
+- [Go](https://golang.org/doc/install) >= 1.19
+
+## Steps
+
+* Create a new directory
+ * ```bash
+ mkdir tf_go_client
+ ```
+* Change directory
+ * ```bash
+ cd tf_go_client
+ ```
+* Creates a **go.mod** file to track the code's dependencies
+ * ```bash
+ go mod init main
+ ```
+* Install the Grid3 Go Client
+ * ```bash
+ go get github.com/threefoldtech/tfgrid-sdk-go/grid-client
+ ```
+
+This will make Grid3 Go Client packages available to you.
+
+## References
+
+For more information, you can read the official [Go documentation](https://go.dev/doc/).
\ No newline at end of file
diff --git a/collections/documentation/developers/go/grid3_go_kubernetes.md b/collections/documentation/developers/go/grid3_go_kubernetes.md
new file mode 100644
index 0000000..ba1bab8
--- /dev/null
+++ b/collections/documentation/developers/go/grid3_go_kubernetes.md
@@ -0,0 +1,120 @@
+ Deploying Kubernetes Clusters
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Example](#example)
+
+***
+
+## Introduction
+
+We show how to deploy a Kubernetes cluster with the Go client.
+
+## Example
+
+```go
+import (
+ "fmt"
+ "net"
+
+ "github.com/threefoldtech/tfgrid-sdk-go/grid-client/deployer"
+ "github.com/threefoldtech/tfgrid-sdk-go/grid-client/workloads"
+ "github.com/threefoldtech/tfgrid-sdk-go/grid-proxy/pkg/types"
+ "github.com/threefoldtech/zos/pkg/gridtypes"
+)
+
+func main() {
+
+ // Create Threefold plugin client
+ tfPluginClient, err := deployer.NewTFPluginClient(mnemonics, "sr25519", network, "", "", "", 0, true)
+
+ // Get a free node to deploy
+ freeMRU := uint64(1)
+ freeSRU := uint64(1)
+ status := "up"
+ filter := types.NodeFilter{
+ FreeMRU: &freeMRU,
+ FreeSRU: &freeSRU,
+ Status: &status,
+ }
+ nodeIDs, err := deployer.FilterNodes(tfPluginClient.GridProxyClient, filter)
+ masterNodeID := uint32(nodeIDs[0].NodeID)
+ workerNodeID1 := uint32(nodeIDs[1].NodeID)
+ workerNodeID2 := uint32(nodeIDs[2].NodeID)
+
+ // Create a new network to deploy
+ network := workloads.ZNet{
+ Name: "newNetwork",
+ Description: "A network to deploy",
+ Nodes: []uint32{masterNodeID, workerNodeID1, workerNodeID2},
+ IPRange: gridtypes.NewIPNet(net.IPNet{
+ IP: net.IPv4(10, 1, 0, 0),
+ Mask: net.CIDRMask(16, 32),
+ }),
+ AddWGAccess: true,
+ }
+
+ // Create master and worker nodes to deploy
+ master := workloads.K8sNode{
+ Name: "master",
+ Node: masterNodeID,
+ DiskSize: 1,
+ CPU: 2,
+ Memory: 1024,
+ Planetary: true,
+ Flist: "https://hub.grid.tf/tf-official-apps/threefoldtech-k3s-latest.flist",
+ }
+
+ worker1 := workloads.K8sNode{
+ Name: "worker1",
+ Node: workerNodeID1,
+ DiskSize: 1,
+ CPU: 2,
+ Memory: 1024,
+ Flist: "https://hub.grid.tf/tf-official-apps/threefoldtech-k3s-latest.flist",
+ }
+
+ worker2 := workloads.K8sNode{
+ Name: "worker2",
+ Node: workerNodeID2,
+ DiskSize: 1,
+ Flist: "https://hub.grid.tf/tf-official-apps/threefoldtech-k3s-latest.flist",
+ CPU: 2,
+ Memory: 1024,
+ }
+
+ k8sCluster := workloads.K8sCluster{
+ Master: &master,
+ Workers: []workloads.K8sNode{worker1, worker2},
+ Token: "tokens",
+ SSHKey: publicKey,
+ NetworkName: network.Name,
+ }
+
+ // Deploy the network first
+ err = tfPluginClient.NetworkDeployer.Deploy(ctx, &network)
+
+ // Deploy the k8s cluster
+ err = tfPluginClient.K8sDeployer.Deploy(ctx, &k8sCluster)
+
+ // Load the k8s cluster
+ k8sClusterObj, err := tfPluginClient.State.LoadK8sFromGrid([]uint32{masterNodeID, workerNodeID1, workerNodeID2}, master.Name)
+
+ // Print master node Yggdrasil IP
+ fmt.Println(k8sClusterObj.Master.YggIP)
+
+ // Cancel the VM deployment
+ err = tfPluginClient.K8sDeployer.Cancel(ctx, &k8sCluster)
+
+ // Cancel the network deployment
+ err = tfPluginClient.NetworkDeployer.Cancel(ctx, &network)
+}
+
+```
+
+You should see an output like this:
+
+```bash
+300:e9c4:9048:57cf:6d98:42c6:a7bf:2e3f
+```
diff --git a/collections/documentation/developers/go/grid3_go_load_client.md b/collections/documentation/developers/go/grid3_go_load_client.md
new file mode 100644
index 0000000..fe61a6d
--- /dev/null
+++ b/collections/documentation/developers/go/grid3_go_load_client.md
@@ -0,0 +1,35 @@
+Load Client
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [TFPluginClient Configuration](#tfpluginclient-configuration)
+- [Creating Client](#creating-client)
+
+***
+
+## Introduction
+
+We cover how to load client using the Go client.
+
+## TFPluginClient Configuration
+
+- mnemonics
+- keyType: can be `ed25519` or `sr25519`
+- network: can be `dev`, `qa`, `test` or `main`
+
+## Creating Client
+
+Import `deployer` package to your project:
+
+```go
+import "github.com/threefoldtech/tfgrid-sdk-go/grid-client/deployer"
+```
+
+Create new Client:
+
+```go
+func main() {
+ client, err := deployer.NewTFPluginClient(mnemonics, keyType, network, "", "", "", 0, true)
+}
+```
diff --git a/collections/documentation/developers/go/grid3_go_qsfs.md b/collections/documentation/developers/go/grid3_go_qsfs.md
new file mode 100644
index 0000000..92a352a
--- /dev/null
+++ b/collections/documentation/developers/go/grid3_go_qsfs.md
@@ -0,0 +1,186 @@
+ Deploying QSFS
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Example](#example)
+
+***
+
+## Introduction
+
+We show how to deploy QSFS workloads with the Go client.
+
+## Example
+
+```go
+import (
+ "context"
+ "fmt"
+ "net"
+
+ "github.com/threefoldtech/tfgrid-sdk-go/grid-client/deployer"
+ "github.com/threefoldtech/tfgrid-sdk-go/grid-client/workloads"
+ "github.com/threefoldtech/tfgrid-sdk-go/grid-proxy/pkg/types"
+ "github.com/threefoldtech/zos/pkg/gridtypes"
+)
+
+func main() {
+
+ // Create Threefold plugin client
+ tfPluginClient, err := deployer.NewTFPluginClient(mnemonics, "sr25519", network, "", "", "", 0, true)
+
+ // Get a free node to deploy
+ freeMRU := uint64(2)
+ freeSRU := uint64(20)
+ status := "up"
+ filter := types.NodeFilter{
+ FreeMRU: &freeMRU,
+ FreeSRU: &freeSRU,
+ Status: &status,
+ }
+ nodeIDs, err := deployer.FilterNodes(tfPluginClient.GridProxyClient, filter)
+ nodeID := uint32(nodeIDs[0].NodeID)
+
+ // Create data and meta ZDBs
+ dataZDBs := []workloads.ZDB{}
+ metaZDBs := []workloads.ZDB{}
+ for i := 1; i <= DataZDBNum; i++ {
+ zdb := workloads.ZDB{
+ Name: "qsfsDataZdb" + strconv.Itoa(i),
+ Password: "password",
+ Public: true,
+ Size: 1,
+ Description: "zdb for testing",
+ Mode: zos.ZDBModeSeq,
+ }
+ dataZDBs = append(dataZDBs, zdb)
+ }
+
+ for i := 1; i <= MetaZDBNum; i++ {
+ zdb := workloads.ZDB{
+ Name: "qsfsMetaZdb" + strconv.Itoa(i),
+ Password: "password",
+ Public: true,
+ Size: 1,
+ Description: "zdb for testing",
+ Mode: zos.ZDBModeUser,
+ }
+ metaZDBs = append(metaZDBs, zdb)
+ }
+
+ // Deploy ZDBs
+ dl1 := workloads.NewDeployment("qsfs", nodeID, "", nil, "", nil, append(dataZDBs, metaZDBs...), nil, nil)
+ err = tfPluginClient.DeploymentDeployer.Deploy(ctx, &dl1)
+
+ // result ZDBs
+ resDataZDBs := []workloads.ZDB{}
+ resMetaZDBs := []workloads.ZDB{}
+ for i := 1; i <= DataZDBNum; i++ {
+ res, err := tfPluginClient.State.LoadZdbFromGrid(nodeID, "qsfsDataZdb"+strconv.Itoa(i), dl1.Name)
+ resDataZDBs = append(resDataZDBs, res)
+ }
+ for i := 1; i <= MetaZDBNum; i++ {
+ res, err := tfPluginClient.State.LoadZdbFromGrid(nodeID, "qsfsMetaZdb"+strconv.Itoa(i), dl1.Name)
+ resMetaZDBs = append(resMetaZDBs, res)
+ }
+
+ // backends
+ dataBackends := []workloads.Backend{}
+ metaBackends := []workloads.Backend{}
+ for i := 0; i < DataZDBNum; i++ {
+ dataBackends = append(dataBackends, workloads.Backend{
+ Address: "[" + resDataZDBs[i].IPs[1] + "]" + ":" + fmt.Sprint(resDataZDBs[i].Port),
+ Namespace: resDataZDBs[i].Namespace,
+ Password: resDataZDBs[i].Password,
+ })
+ }
+ for i := 0; i < MetaZDBNum; i++ {
+ metaBackends = append(metaBackends, workloads.Backend{
+ Address: "[" + resMetaZDBs[i].IPs[1] + "]" + ":" + fmt.Sprint(resMetaZDBs[i].Port),
+ Namespace: resMetaZDBs[i].Namespace,
+ Password: resMetaZDBs[i].Password,
+ })
+ }
+
+ // Create a new qsfs to deploy
+ qsfs := workloads.QSFS{
+ Name: "qsfs",
+ Description: "qsfs for testing",
+ Cache: 1024,
+ MinimalShards: 2,
+ ExpectedShards: 4,
+ RedundantGroups: 0,
+ RedundantNodes: 0,
+ MaxZDBDataDirSize: 512,
+ EncryptionAlgorithm: "AES",
+ EncryptionKey: "4d778ba3216e4da4231540c92a55f06157cabba802f9b68fb0f78375d2e825af",
+ CompressionAlgorithm: "snappy",
+ Groups: workloads.Groups{{Backends: dataBackends}},
+ Metadata: workloads.Metadata{
+ Type: "zdb",
+ Prefix: "test",
+ EncryptionAlgorithm: "AES",
+ EncryptionKey: "4d778ba3216e4da4231540c92a55f06157cabba802f9b68fb0f78375d2e825af",
+ Backends: metaBackends,
+ },
+ }
+
+ // Create a new network to deploy
+ network := workloads.ZNet{
+ Name: "newNetwork",
+ Description: "A network to deploy",
+ Nodes: []uint32{nodeID},
+ IPRange: gridtypes.NewIPNet(net.IPNet{
+ IP: net.IPv4(10, 1, 0, 0),
+ Mask: net.CIDRMask(16, 32),
+ }),
+ AddWGAccess: true,
+ }
+
+ vm := workloads.VM{
+ Name: "vm",
+ Flist: "https://hub.grid.tf/tf-official-apps/base:latest.flist",
+ CPU: 2,
+ Planetary: true,
+ Memory: 1024,
+ Entrypoint: "/sbin/zinit init",
+ EnvVars: map[string]string{
+ "SSH_KEY": publicKey,
+ },
+ Mounts: []workloads.Mount{
+ {DiskName: qsfs.Name, MountPoint: "/qsfs"},
+ },
+ NetworkName: network.Name,
+ }
+
+ // Deploy the network first
+ err = tfPluginClient.NetworkDeployer.Deploy(ctx, &network)
+
+ // Deploy the VM/QSFS deployment
+ dl2 := workloads.NewDeployment("qsfs", nodeID, "", nil, network.Name, nil, append(dataZDBs, metaZDBs...), []workloads.VM{vm}, []workloads.QSFS{qsfs})
+ err = tfPluginClient.DeploymentDeployer.Deploy(ctx, &dl2)
+
+ // Load the QSFS using the state loader
+ qsfsObj, err := tfPluginClient.State.LoadQSFSFromGrid(nodeID, qsfs.Name, dl2.Name)
+
+ // Load the VM using the state loader
+ vmObj, err := tfPluginClient.State.LoadVMFromGrid(nodeID, vm.Name, dl2.Name)
+
+ // Print the VM Yggdrasil IP
+ fmt.Println(vmObj.YggIP)
+
+ // Cancel the VM,QSFS deployment
+ err = tfPluginClient.DeploymentDeployer.Cancel(ctx, &dl1)
+ err = tfPluginClient.DeploymentDeployer.Cancel(ctx, &dl2)
+
+ // Cancel the network deployment
+ err = tfPluginClient.NetworkDeployer.Cancel(ctx, &network)
+}
+```
+
+Running this code should result in a VM with QSFS deployed on an available node and get an output like this:
+
+```bash
+Yggdrasil IP: 300:e9c4:9048:57cf:6d98:42c6:a7bf:2e3f
+```
diff --git a/collections/documentation/developers/go/grid3_go_readme.md b/collections/documentation/developers/go/grid3_go_readme.md
new file mode 100644
index 0000000..cf24ada
--- /dev/null
+++ b/collections/documentation/developers/go/grid3_go_readme.md
@@ -0,0 +1,17 @@
+# Grid Go Client
+
+Grid Go Client is a Go client created to interact and develop on Threefold Grid using Go language.
+
+Please make sure to check the [basics](../../system_administrators/getstarted/tfgrid3_getstarted.md) before continuing.
+
+ Table of Contents
+
+- [Installation](../go/grid3_go_installation.md)
+- [Loading Client](../go/grid3_go_load_client.md)
+- [Deploy a VM](../go/grid3_go_vm.md)
+- [Deploy a VM with GPU](../go/grid3_go_vm_with_gpu.md)
+- [Deploy Multiple VMs](../go/grid3_go_vms.md)
+- [Deploy Gateways](../go/grid3_go_gateways.md)
+- [Deploy Kubernetes](../go/grid3_go_kubernetes.md)
+- [Deploy a QSFS](../go/grid3_go_qsfs.md)
+- [GPU Support](../go/grid3_go_gpu_support.md)
diff --git a/collections/documentation/developers/go/grid3_go_vm.md b/collections/documentation/developers/go/grid3_go_vm.md
new file mode 100644
index 0000000..c164114
--- /dev/null
+++ b/collections/documentation/developers/go/grid3_go_vm.md
@@ -0,0 +1,99 @@
+ Deploying a VM
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Example](#example)
+
+***
+
+## Introduction
+
+We show how to deploy a VM with the Go client.
+
+## Example
+
+```go
+import (
+ "context"
+ "fmt"
+ "net"
+
+ "github.com/threefoldtech/tfgrid-sdk-go/grid-client/deployer"
+ "github.com/threefoldtech/tfgrid-sdk-go/grid-client/workloads"
+ "github.com/threefoldtech/tfgrid-sdk-go/grid-proxy/pkg/types"
+ "github.com/threefoldtech/zos/pkg/gridtypes"
+)
+
+func main() {
+
+ // Create Threefold plugin client
+ tfPluginClient, err := deployer.NewTFPluginClient(mnemonics, keyType, network, "", "", "", 0, true)
+
+ // Get a free node to deploy
+ freeMRU := uint64(2)
+ freeSRU := uint64(20)
+ status := "up"
+ filter := types.NodeFilter{
+ FreeMRU: &freeMRU,
+ FreeSRU: &freeSRU,
+ Status: &status,
+ }
+ nodeIDs, err := deployer.FilterNodes(tfPluginClient.GridProxyClient, filter)
+ nodeID := uint32(nodeIDs[0].NodeID)
+
+ // Create a new network to deploy
+ network := workloads.ZNet{
+ Name: "newNetwork",
+ Description: "A network to deploy",
+ Nodes: []uint32{nodeID},
+ IPRange: gridtypes.NewIPNet(net.IPNet{
+ IP: net.IPv4(10, 1, 0, 0),
+ Mask: net.CIDRMask(16, 32),
+ }),
+ AddWGAccess: true,
+ }
+
+ // Create a new VM to deploy
+ vm := workloads.VM{
+ Name: "vm",
+ Flist: "https://hub.grid.tf/tf-official-apps/base:latest.flist",
+ CPU: 2,
+ PublicIP: true,
+ Planetary: true,
+ Memory: 1024,
+ RootfsSize: 20 * 1024,
+ Entrypoint: "/sbin/zinit init",
+ EnvVars: map[string]string{
+ "SSH_KEY": publicKey,
+ },
+ IP: "10.20.2.5",
+ NetworkName: network.Name,
+ }
+
+ // Deploy the network first
+ err = tfPluginClient.NetworkDeployer.Deploy(ctx, &network)
+
+ // Deploy the VM deployment
+ dl := workloads.NewDeployment("vm", nodeID, "", nil, network.Name, nil, nil, []workloads.VM{vm}, nil)
+ err = tfPluginClient.DeploymentDeployer.Deploy(ctx, &dl)
+
+ // Load the VM using the state loader
+ vmObj, err := tfPluginClient.State.LoadVMFromGrid(nodeID, vm.Name, dl.Name)
+
+ // Print the VM Yggdrasil IP
+ fmt.Println(vmObj.YggIP)
+
+ // Cancel the VM deployment
+ err = tfPluginClient.DeploymentDeployer.Cancel(ctx, &dl)
+
+ // Cancel the network deployment
+ err = tfPluginClient.NetworkDeployer.Cancel(ctx, &network)
+}
+```
+
+Running this code should result in a VM deployed on an available node and get an output like this:
+
+```bash
+300:e9c4:9048:57cf:6d98:42c6:a7bf:2e3f
+```
diff --git a/collections/documentation/developers/go/grid3_go_vm_with_gpu.md b/collections/documentation/developers/go/grid3_go_vm_with_gpu.md
new file mode 100644
index 0000000..d635800
--- /dev/null
+++ b/collections/documentation/developers/go/grid3_go_vm_with_gpu.md
@@ -0,0 +1,121 @@
+ Deploy a VM with GPU
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Example](#example)
+
+***
+
+## Introduction
+
+In this section, we explore how to deploy a virtual machine equipped with GPU. We deploy the VM using Go. The VM will be deployed on a 3Node with an available GPU.
+
+
+
+## Example
+
+```go
+import (
+ "context"
+ "fmt"
+ "net"
+
+ "github.com/threefoldtech/tfgrid-sdk-go/grid-client/deployer"
+ "github.com/threefoldtech/tfgrid-sdk-go/grid-client/workloads"
+ "github.com/threefoldtech/tfgrid-sdk-go/grid-proxy/pkg/types"
+ "github.com/threefoldtech/zos/pkg/gridtypes"
+)
+
+func main() {
+
+ // Create Threefold plugin client
+ tfPluginClient, err := deployer.NewTFPluginClient(mnemonics, "sr25519", network, "", "", "", 0, true)
+
+ // Get a free node to deploy
+ freeMRU := uint64(2)
+ freeSRU := uint64(20)
+ status := "up"
+ trueVal := true
+
+ twinID := uint64(tfPluginClient.TwinID)
+ filter := types.NodeFilter{
+ FreeMRU: &freeMRU,
+ FreeSRU: &freeSRU,
+ Status: &status,
+ RentedBy: &twinID,
+ HasGPU: &trueVal,
+ }
+ nodeIDs, err := deployer.FilterNodes(tfPluginClient.GridProxyClient, filter)
+ nodeID := uint32(nodeIDs[0].NodeID)
+
+ // Get the available gpus on the node
+ nodeClient, err := tfPluginClient.NcPool.GetNodeClient(tfPluginClient.SubstrateConn, nodeID)
+ gpus, err := nodeClient.GPUs(ctx)
+
+ // Create a new network to deploy
+ network := workloads.ZNet{
+ Name: "newNetwork",
+ Description: "A network to deploy",
+ Nodes: []uint32{nodeID},
+ IPRange: gridtypes.NewIPNet(net.IPNet{
+ IP: net.IPv4(10, 1, 0, 0),
+ Mask: net.CIDRMask(16, 32),
+ }),
+ AddWGAccess: true,
+ }
+
+ // Create a new disk to deploy
+ disk := workloads.Disk{
+ Name: "gpuDisk",
+ SizeGB: 20,
+ }
+
+ // Create a new VM to deploy
+ vm := workloads.VM{
+ Name: "vm",
+ Flist: "https://hub.grid.tf/tf-official-apps/base:latest.flist",
+ CPU: 2,
+ PublicIP: true,
+ Planetary: true,
+ // Insert your GPUs' IDs here
+ GPUs: []zos.GPU{zos.GPU(gpus[0].ID)},
+ Memory: 1024,
+ RootfsSize: 20 * 1024,
+ Entrypoint: "/sbin/zinit init",
+ EnvVars: map[string]string{
+ "SSH_KEY": publicKey,
+ },
+ Mounts: []workloads.Mount{
+ {DiskName: disk.Name, MountPoint: "/data"},
+ },
+ IP: "10.20.2.5",
+ NetworkName: network.Name,
+ }
+
+ // Deploy the network first
+ err = tfPluginClient.NetworkDeployer.Deploy(ctx, &network)
+
+ // Deploy the VM deployment
+ dl := workloads.NewDeployment("gpu", nodeID, "", nil, network.Name, []workloads.Disk{disk}, nil, []workloads.VM{vm}, nil)
+ err = tfPluginClient.DeploymentDeployer.Deploy(ctx, &dl)
+
+ // Load the VM using the state loader
+ vmObj, err := tfPluginClient.State.LoadVMFromGrid(nodeID, vm.Name, dl.Name)
+
+ // Print the VM Yggdrasil IP
+ fmt.Println(vmObj.YggIP)
+
+ // Cancel the VM deployment
+ err = tfPluginClient.DeploymentDeployer.Cancel(ctx, &dl)
+
+ // Cancel the network deployment
+ err = tfPluginClient.NetworkDeployer.Cancel(ctx, &network)
+}
+```
+
+Running this code should result in a VM with a GPU deployed on an available node. The output should look like this:
+
+```bash
+Yggdrasil IP: 300:e9c4:9048:57cf:6d98:42c6:a7bf:2e3f
+```
\ No newline at end of file
diff --git a/collections/documentation/developers/go/grid3_go_vms.md b/collections/documentation/developers/go/grid3_go_vms.md
new file mode 100644
index 0000000..f78aaae
--- /dev/null
+++ b/collections/documentation/developers/go/grid3_go_vms.md
@@ -0,0 +1,125 @@
+ Deploying Multiple VMs
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Example](#example)
+
+***
+
+## Introduction
+
+We show how to deploy multiple VMs with the Go client.
+
+## Example
+
+```go
+import (
+ "context"
+ "fmt"
+ "net"
+
+ "github.com/threefoldtech/tfgrid-sdk-go/grid-client/deployer"
+ "github.com/threefoldtech/tfgrid-sdk-go/grid-client/workloads"
+ "github.com/threefoldtech/tfgrid-sdk-go/grid-proxy/pkg/types"
+ "github.com/threefoldtech/zos/pkg/gridtypes"
+)
+
+func main() {
+
+ // Create Threefold plugin client
+ tfPluginClient, err := deployer.NewTFPluginClient(mnemonics, "sr25519", network, "", "", "", 0, true)
+
+ // Get a free node to deploy
+ freeMRU := uint64(2)
+ freeSRU := uint64(2)
+ status := "up"
+ filter := types.NodeFilter {
+ FreeMRU: &freeMRU,
+ FreeSRU: &freeSRU,
+ Status: &status,
+ }
+ nodeIDs, err := deployer.FilterNodes(tfPluginClient.GridProxyClient, filter)
+ nodeID1 := uint32(nodeIDs[0].NodeID)
+ nodeID2 := uint32(nodeIDs[1].NodeID)
+
+ // Create a new network to deploy
+ network := workloads.ZNet{
+ Name: "newNetwork",
+ Description: "A network to deploy",
+ Nodes: []uint32{nodeID1, nodeID2},
+ IPRange: gridtypes.NewIPNet(net.IPNet{
+ IP: net.IPv4(10, 1, 0, 0),
+ Mask: net.CIDRMask(16, 32),
+ }),
+ AddWGAccess: true,
+ }
+
+ // Create new VMs to deploy
+ vm1 := workloads.VM{
+ Name: "vm1",
+ Flist: "https://hub.grid.tf/tf-official-apps/base:latest.flist",
+ CPU: 2,
+ PublicIP: true,
+ Planetary: true,
+ Memory: 1024,
+ RootfsSize: 20 * 1024,
+ Entrypoint: "/sbin/zinit init",
+ EnvVars: map[string]string{
+ "SSH_KEY": publicKey,
+ },
+ IP: "10.20.2.5",
+ NetworkName: network.Name,
+ }
+ vm2 := workloads.VM{
+ Name: "vm2",
+ Flist: "https://hub.grid.tf/tf-official-apps/base:latest.flist",
+ CPU: 2,
+ PublicIP: true,
+ Planetary: true,
+ Memory: 1024,
+ RootfsSize: 20 * 1024,
+ Entrypoint: "/sbin/zinit init",
+ EnvVars: map[string]string{
+ "SSH_KEY": publicKey,
+ },
+ IP: "10.20.2.6",
+ NetworkName: network.Name,
+ }
+
+ // Deploy the network first
+ err = tfPluginClient.NetworkDeployer.Deploy(ctx, &network)
+
+ // Load the network using the state loader
+ // this loader should load the deployment as json then convert it to a deployment go object with workloads inside it
+ networkObj, err := tfPluginClient.State.LoadNetworkFromGrid(network.Name)
+
+ // Deploy the VM deployments
+ dl1 := workloads.NewDeployment("vm1", nodeID1, "", nil, network.Name, nil, nil, []workloads.VM{vm1}, nil)
+ dl2 := workloads.NewDeployment("vm2", nodeID2, "", nil, network.Name, nil, nil, []workloads.VM{vm2}, nil)
+ err = tfPluginClient.DeploymentDeployer.BatchDeploy(ctx, []*workloads.Deployment{&dl1, &dl2})
+
+ // Load the VMs using the state loader
+ vmObj1, err := tfPluginClient.State.LoadVMFromGrid(nodeID1, vm1.Name, dl1.Name)
+ vmObj2, err := tfPluginClient.State.LoadVMFromGrid(nodeID2, vm2.Name, dl2.Name)
+
+ // Print the VMs Yggdrasil IP
+ fmt.Println(vmObj1.YggIP)
+ fmt.Println(vmObj2.YggIP)
+
+ // Cancel the VM deployments
+ err = tfPluginClient.DeploymentDeployer.Cancel(ctx, &dl1)
+ err = tfPluginClient.DeploymentDeployer.Cancel(ctx, &dl2)
+
+ // Cancel the network
+ err = tfPluginClient.NetworkDeployer.Cancel(ctx, &network)
+}
+
+```
+
+Running this code should result in two VMs deployed on two separate nodes while being on the same network and you should see an output like this:
+
+```bash
+300:e9c4:9048:57cf:f4e0:2343:f891:6037
+300:e9c4:9048:57cf:6d98:42c6:a7bf:2e3f
+```
diff --git a/collections/documentation/developers/grid_deployment/grid_deployment.md b/collections/documentation/developers/grid_deployment/grid_deployment.md
new file mode 100644
index 0000000..14fa506
--- /dev/null
+++ b/collections/documentation/developers/grid_deployment/grid_deployment.md
@@ -0,0 +1,9 @@
+# Grid Deployment
+
+The TFGrid whole source code is open-source and instances of the grid can be deployed by anyone thanks to the distribution of daily grid snapshots of the complete ThreeFold Grid stacks.
+
+## Table of Contents
+
+- [TFGrid Stacks](./tfgrid_stacks.md)
+- [Full VM Grid Deployment](./grid_deployment_full_vm.md)
+- [Grid Snapshots](./snapshots.md)
\ No newline at end of file
diff --git a/collections/documentation/developers/grid_deployment/grid_deployment_full_vm.md b/collections/documentation/developers/grid_deployment/grid_deployment_full_vm.md
new file mode 100644
index 0000000..aee260a
--- /dev/null
+++ b/collections/documentation/developers/grid_deployment/grid_deployment_full_vm.md
@@ -0,0 +1,152 @@
+ Grid Deployment on a Full VM
+Table of Contents
+
+
+- [Introduction](#introduction)
+- [Prerequisites](#prerequisites)
+- [DNS Settings](#dns-settings)
+ - [DNS Verification](#dns-verification)
+- [Prepare the VM](#prepare-the-vm)
+- [Set the Firewall](#set-the-firewall)
+- [Launch the Script](#launch-the-script)
+- [Access the Grid Services](#access-the-grid-services)
+- [Manual Commands](#manual-commands)
+- [Update the Deployment](#update-the-deployment)
+
+***
+
+## Introduction
+
+We present the steps to deploy a network instance of the TFGrid on a full VM.
+
+For this guide, we will be deploying a mainnet instance. While the steps are similar for testnet and devnet, you will have to adjust your deployment depending on which network you use.
+
+## Prerequisites
+
+For this guide, you will need to deploy a full VM on the ThreeFold Grid with at least the following minimum specs:
+
+- IPv4
+- IPv6
+- 32GB of RAM
+- 1000 GB of SSD
+- 8 vcores
+
+After deploying the full VM, take note of the IPv4 and IPv6 addresses to properly set the DNS records and then SSH into the VM.
+
+## DNS Settings
+
+You need to set an A record for the IPv4 address and an AAAA record for the IPv6 address with a wildcard subdomain.
+
+The following table explicitly shows how to set the A and AAAA records for your domain.
+
+| Type | Host | Value |
+| ---- | ---- | -------------- |
+| A | \* | |
+| AAAA | \* | |
+
+
+### DNS Verification
+
+You can use tools such as [DNSChecker](https://dnschecker.org/) or [dig](https://linux.die.net/man/1/dig) on a terminal to check if the DNS propagadation is complete.
+
+## Prepare the VM
+
+- Download the ThreeFold Tech `grid_deployment` repository
+ ```
+ git clone https://github.com/threefoldtech/grid_deployment
+ cd grid_deployment/docker-compose/mainnet
+ ```
+- Generate a TFChain node key with `subkey`
+ ```
+ echo .subkey_mainnet >> .gitignore
+ ../subkey generate-node-key > .nodekey_mainnet
+ cat .nodekey_mainnet
+ ```
+- Create and the set environment variables file
+ ```
+ cp .secrets.env-example .secrets.env
+ ```
+- Adjust the environment file
+ ```
+ nano .secrets.env
+ ```
+- To adjust the `.secrets.env` file, take into account the following:
+ - **DOMAIN**="example.com"
+ - Write your own domain
+ - **TFCHAIN_NODE_KEY**="abc123"
+ - Write the output of the command `cat .nodekey_mainnet`
+ - **ACTIVATION_SERVICE_MNEMONIC**="word1 word2 ... word24"
+ - Write the seed phrase of an account on mainnet with at least 10 TFT in the wallet
+ - **GRID_PROXY_MNEMONIC**="word1 word2 ... word24"
+ - Write the seed phrase of an account on mainnet with at least 10 TFT in the wallet and a registered twin ID\*
+
+> \*Note: If you've created an account using the ThreeFold Dashboard on mainnet, the twin ID is automatically registered.
+
+## Set the Firewall
+
+You can use UFW to set the firewall:
+
+```
+ufw allow 80/tcp
+ufw allow 443/tcp
+ufw allow 30333/tcp
+ufw allow 22/tcp
+ufw enable
+ufw status
+```
+
+## Launch the Script
+
+Once you've prepared the VM, you can simply run the script to install the grid stack and deploy it online.
+
+```
+sh install_grid_bknd.sh
+```
+
+This will take some time since you are downloading the whole mainnet grid snapshots.
+
+## Access the Grid Services
+
+Once you've deployed the grid stack online, you can access the different grid services by usual the usual subdomains:
+
+```
+dashboard.your.domain
+metrics.your.domain
+tfchain.your.domain
+graphql.your.domain
+relay.your.domain
+gridproxy.your.domain
+activation.your.domain
+stats.your.domain
+```
+
+## Manual Commands
+
+Once you've run the install script, you can deploy manually the grid stack with the following command:
+
+```
+docker compose --env-file .secrets.env --env-file .env up -d
+```
+
+You can also check if the environment variables are properly set:
+
+```
+docker compose --env-file .secrets.env --env-file .env config
+```
+
+If you want to see the output during deployment, remove `-d` in the command above as follows:
+
+```
+docker compose --env-file .secrets.env --env-file .env up
+```
+
+This can be helpful to troubleshoot errors.
+
+## Update the Deployment
+
+Go into the folder of the proper network, e.g. mainnet, and run the following commands:
+
+```
+git pull -r
+docker compose --env-file .secrets.env --env-file .env up -d
+```
\ No newline at end of file
diff --git a/collections/documentation/developers/grid_deployment/snapshots.md b/collections/documentation/developers/grid_deployment/snapshots.md
new file mode 100644
index 0000000..2b7a7d4
--- /dev/null
+++ b/collections/documentation/developers/grid_deployment/snapshots.md
@@ -0,0 +1,196 @@
+Snapshots for Grid Backend Services
+Table of Contents
+
+- [Introduction](#introduction)
+- [Services](#services)
+- [ThreeFold Public Snapshots](#threefold-public-snapshots)
+- [Deploy the Services with Scripts](#deploy-the-services-with-scripts)
+ - [Create the Snapshots](#create-the-snapshots)
+ - [Start All the Services](#start-all-the-services)
+ - [Stop All the Services](#stop-all-the-services)
+- [Expose the Snapshots with Rsync](#expose-the-snapshots-with-rsync)
+ - [Create the Service Configuration File](#create-the-service-configuration-file)
+ - [Start the Service](#start-the-service)
+
+***
+
+## Introduction
+
+To facilitate deploying grid backend services, we provide snapshots to significantly reduce sync time. This can be setup anywhere from scratch. Once all services are synced, one can use the scripts to create snapshots automatically.
+
+To learn how to deploy your own grid stack, read [this section](./grid_deployment_full_vm.md).
+
+## Services
+
+There are 3 grid backend services that collect enough data to justify creating snapshots:
+
+- ThreeFold blockchain - TFChain
+- Graphql - Indexer
+- Graphql - Processor
+
+## ThreeFold Public Snapshots
+
+ThreeFold hosts all available snapshots at: [https://bknd.snapshot.grid.tf/](https://bknd.snapshot.grid.tf/). Those snapshots can be downloaded with rsync:
+
+- Mainnet:
+ ```
+ rsync -Lv --progress --partial rsync://bknd.snapshot.grid.tf:34873/gridsnapshots/tfchain-mainnet-latest.tar.gz .
+ rsync -Lv --progress --partial rsync://bknd.snapshot.grid.tf:34873/gridsnapshots/indexer-mainnet-latest.tar.gz .
+ rsync -Lv --progress --partial rsync://bknd.snapshot.grid.tf:34873/gridsnapshots/processor-mainnet-latest.tar.gz .
+ ```
+- Testnet:
+ ```
+ rsync -Lv --progress --partial rsync://bknd.snapshot.grid.tf:34873/gridsnapshotstest/tfchain-testnet-latest.tar.gz .
+ rsync -Lv --progress --partial rsync://bknd.snapshot.grid.tf:34873/gridsnapshotstest/indexer-testnet-latest.tar.gz .
+ rsync -Lv --progress --partial rsync://bknd.snapshot.grid.tf:34873/gridsnapshotstest/processor-testnet-latest.tar.gz .
+ ```
+- Devnet:
+ ```
+ rsync -Lv --progress --partial rsync://bknd.snapshot.grid.tf:34873/gridsnapshotsdev/tfchain-devnet-latest.tar.gz .
+ rsync -Lv --progress --partial rsync://bknd.snapshot.grid.tf:34873/gridsnapshotsdev/indexer-devnet-latest.tar.gz .
+ rsync -Lv --progress --partial rsync://bknd.snapshot.grid.tf:34873/gridsnapshotsdev/processor-devnet-latest.tar.gz .
+ ```
+
+## Deploy the Services with Scripts
+
+You can deploy the 3 individual services using known methods such as [Docker](../../system_administrators/computer_it_basics/docker_basics.md). To facilitate the process, scripts are provided that run the necessary docker commands.
+
+The first script creates the snapshots, while the second and third scripts serve to start and stop all services.
+
+You can use the start script to start all services and then set a cron job to execute periodically the snapshot creation script. This will ensure that you always have the latest version available on your server.
+
+### Create the Snapshots
+
+You can set a cron job to execute a script running rsync to create the snapshots and generate logs at a given interval.
+
+- First download the script.
+ - Main net
+ ```
+ wget https://github.com/threefoldtech/grid_deployment/blob/development/grid-snapshots/mainnet/create_snapshot.sh
+ ```
+ - Test net
+ ```
+ wget https://github.com/threefoldtech/grid_deployment/blob/development/grid-snapshots/testnet/create_snapshot.sh
+ ```
+ - Dev net
+ ```
+ wget https://github.com/threefoldtech/grid_deployment/blob/development/grid-snapshots/devnet/create_snapshot.sh
+ ```
+- Set the permissions of the script
+ ```
+ chmod +x create_snapshot.sh
+ ```
+- Make sure to a adjust the snapshot creation script for your specific deployment
+- Set a cron job
+ ```
+ crontab -e
+ ```
+ - Here is an example of a cron job where we execute the script every day at 1 AM and send the logs to `/var/log/snapshots/snapshots-cron.log`.
+ ```sh
+ 0 1 * * * sh /opt/snapshots/create-snapshot.sh > /var/log/snapshots/snapshots-cron.log 2>&1
+ ```
+
+### Start All the Services
+
+You can start all services by running the provided scripts.
+
+- Download the script.
+ - Main net
+ ```
+ wget https://github.com/threefoldtech/grid_deployment/blob/development/grid-snapshots/mainnet/startall.sh
+ ```
+ - Test net
+ ```
+ wget https://github.com/threefoldtech/grid_deployment/blob/development/grid-snapshots/testnet/startall.sh
+ ```
+ - Dev net
+ ```
+ wget https://github.com/threefoldtech/grid_deployment/blob/development/grid-snapshots/devnet/startall.sh
+ ```
+- Set the permissions of the script
+ ```
+ chmod +x startall.sh
+ ```
+- Run the script to start all services via docker engine.
+ ```
+ ./startall.sh
+ ```
+
+### Stop All the Services
+
+You can stop all services by running the provided scripts.
+
+- Download the script.
+ - Main net
+ ```
+ wget https://github.com/threefoldtech/grid_deployment/blob/development/grid-snapshots/mainnet/stopall.sh
+ ```
+ - Test net
+ ```
+ wget https://github.com/threefoldtech/grid_deployment/blob/development/grid-snapshots/testnet/stopall.sh
+ ```
+ - Dev net
+ ```
+ wget https://github.com/threefoldtech/grid_deployment/blob/development/grid-snapshots/devnet/stopall.sh
+ ```
+- Set the permissions of the script
+ ```
+ chmod +x stopall.sh
+ ```
+- Run the script to stop all services via docker engine.
+ ```
+ ./stopall.sh
+ ```
+
+## Expose the Snapshots with Rsync
+
+We use rsync with a systemd service to expose the snapshots to the community.
+
+### Create the Service Configuration File
+
+To setup a public rsync server, create and edit the following file:
+
+`/etc/rsyncd.conf`
+
+```sh
+pid file = /var/run/rsyncd.pid
+lock file = /var/run/rsync.lock
+log file = /var/log/rsync.log
+port = 34873
+max connections = 20
+exclude = lost+found/
+transfer logging = yes
+use chroot = yes
+reverse lookup = no
+
+[gridsnapshots]
+path = /storage/rsync-public/mainnet
+comment = THREEFOLD GRID MAINNET SNAPSHOTS
+read only = true
+timeout = 300
+list = false
+
+[gridsnapshotstest]
+path = /storage/rsync-public/testnet
+comment = THREEFOLD GRID TESTNET SNAPSHOTS
+read only = true
+timeout = 300
+list = false
+
+[gridsnapshotsdev]
+path = /storage/rsync-public/devnet
+comment = THREEFOLD GRID DEVNET SNAPSHOTS
+read only = true
+timeout = 300
+list = false
+```
+
+### Start the Service
+
+Start and enable via systemd:
+
+```sh
+systemctl start rsync
+systemctl enable rsync
+systemctl status rsync
+```
diff --git a/collections/documentation/developers/grid_deployment/snapshots_archive.md b/collections/documentation/developers/grid_deployment/snapshots_archive.md
new file mode 100644
index 0000000..eac3ff7
--- /dev/null
+++ b/collections/documentation/developers/grid_deployment/snapshots_archive.md
@@ -0,0 +1,206 @@
+Snapshots for Grid Backend Services
+Table of Contents
+
+- [Introduction](#introduction)
+- [Prerequisites](#prerequisites)
+- [Services](#services)
+- [ThreeFold Public Snapshots](#threefold-public-snapshots)
+- [Deploy the Services with Scripts](#deploy-the-services-with-scripts)
+ - [Start All the Services](#start-all-the-services)
+ - [Stop All the Services](#stop-all-the-services)
+ - [Create the Snapshots](#create-the-snapshots)
+- [Expose the Snapshots with Rsync](#expose-the-snapshots-with-rsync)
+ - [Create the Service Configuration File](#create-the-service-configuration-file)
+ - [Start the Service](#start-the-service)
+
+***
+
+## Introduction
+
+To facilitate deploying grid backend services, we provide snapshots to significantly reduce sync time. This can be setup anywhere from scratch. Once all services are synced, one can use the scripts to create snapshots automatically.
+
+## Prerequisites
+
+There are a few prerequisites to properly run the ThreeFold services.
+
+- [Docker engine](../computer_it_basics/docker_basics.md#install-docker-desktop-and-docker-engine)
+- [Rsync](../computer_it_basics/file_transfer.md#rsync)
+
+## Services
+
+There are 3 grid backend services that collect enough data to justify creating snapshots:
+
+- ThreeFold blockchain - TFChain
+- Graphql - Indexer
+- Graphql - Processor
+
+## ThreeFold Public Snapshots
+
+ThreeFold hosts all available snapshots at: [https://bknd.snapshot.grid.tf/](https://bknd.snapshot.grid.tf/). Those snapshots can be downloaded with rsync:
+
+- Mainnet:
+ ```
+ rsync -Lv --progress --partial rsync://bknd.snapshot.grid.tf:34873/gridsnapshots/tfchain-mainnet-latest.tar.gz .
+ rsync -Lv --progress --partial rsync://bknd.snapshot.grid.tf:34873/gridsnapshots/indexer-mainnet-latest.tar.gz .
+ rsync -Lv --progress --partial rsync://bknd.snapshot.grid.tf:34873/gridsnapshots/processor-mainnet-latest.tar.gz .
+ ```
+- Testnet:
+ ```
+ rsync -Lv --progress --partial rsync://bknd.snapshot.grid.tf:34873/gridsnapshotstest/tfchain-testnet-latest.tar.gz .
+ rsync -Lv --progress --partial rsync://bknd.snapshot.grid.tf:34873/gridsnapshotstest/indexer-testnet-latest.tar.gz .
+ rsync -Lv --progress --partial rsync://bknd.snapshot.grid.tf:34873/gridsnapshotstest/processor-testnet-latest.tar.gz .
+ ```
+- Devnet:
+ ```
+ rsync -Lv --progress --partial rsync://bknd.snapshot.grid.tf:34873/gridsnapshotsdev/tfchain-devnet-latest.tar.gz .
+ rsync -Lv --progress --partial rsync://bknd.snapshot.grid.tf:34873/gridsnapshotsdev/indexer-devnet-latest.tar.gz .
+ rsync -Lv --progress --partial rsync://bknd.snapshot.grid.tf:34873/gridsnapshotsdev/processor-devnet-latest.tar.gz .
+ ```
+
+Let's now see how to use those snapshots to run the services via scripts.
+
+## Deploy the Services with Scripts
+
+You can deploy the 3 individual services using known methods such as [Docker](https://manual.grid.tf/computer_it_basics/docker_basics.html). To facilitate the process, scripts are provided that run the necessary docker commands.
+
+The first script creates the snapshots, while the second and third scripts serve to start and stop all services.
+
+You can use the start script to start all services and then set a cron job to execute periodically the snapshot creation script. This will ensure that you always have the latest version available on your server.
+
+### Start All the Services
+
+You can start all services by running the provided scripts.
+
+- Download the script.
+ - Main net
+ ```
+ wget https://github.com/threefoldtech/grid_deployment/blob/development/grid-snapshots/mainnet/startall.sh
+ ```
+ - Test net
+ ```
+ wget https://github.com/threefoldtech/grid_deployment/blob/development/grid-snapshots/testnet/startall.sh
+ ```
+ - Dev net
+ ```
+ wget https://github.com/threefoldtech/grid_deployment/blob/development/grid-snapshots/devnet/startall.sh
+ ```
+- Set the permissions of the script
+ ```
+ chmod +x startall.sh
+ ```
+- Run the script to start all services via docker engine.
+ ```
+ ./startall.sh
+ ```
+
+### Stop All the Services
+
+You can stop all services by running the provided scripts.
+
+- Download the script.
+ - Main net
+ ```
+ wget https://github.com/threefoldtech/grid_deployment/blob/development/grid-snapshots/mainnet/stopall.sh
+ ```
+ - Test net
+ ```
+ wget https://github.com/threefoldtech/grid_deployment/blob/development/grid-snapshots/testnet/stopall.sh
+ ```
+ - Dev net
+ ```
+ wget https://github.com/threefoldtech/grid_deployment/blob/development/grid-snapshots/devnet/stopall.sh
+ ```
+- Set the permissions of the script
+ ```
+ chmod +x stopall.sh
+ ```
+- Run the script to stop all services via docker engine.
+ ```
+ ./stopall.sh
+ ```
+
+### Create the Snapshots
+
+You can set a cron job to execute a script running rsync to create the snapshots and generate logs at a given interval.
+
+- First download the script.
+ - Main net
+ ```
+ wget https://github.com/threefoldtech/grid_deployment/blob/development/grid-snapshots/mainnet/create_snapshot.sh
+ ```
+ - Test net
+ ```
+ wget https://github.com/threefoldtech/grid_deployment/blob/development/grid-snapshots/testnet/create_snapshot.sh
+ ```
+ - Dev net
+ ```
+ wget https://github.com/threefoldtech/grid_deployment/blob/development/grid-snapshots/devnet/create_snapshot.sh
+ ```
+- Set the permissions of the script
+ ```
+ chmod +x create_snapshot.sh
+ ```
+- Make sure to a adjust the snapshot creation script for your specific deployment
+- Set a cron job
+ ```
+ crontab -e
+ ```
+ - Here is an example of a cron job where we execute the script every day at 1 AM and send the logs to `/var/log/snapshots/snapshots-cron.log`.
+ ```sh
+ 0 1 * * * sh /opt/snapshots/create-snapshot.sh > /var/log/snapshots/snapshots-cron.log 2>&1
+ ```
+
+## Expose the Snapshots with Rsync
+
+We use rsync with a systemd service to expose the snapshots to the community.
+
+### Create the Service Configuration File
+
+To setup a public rsync server, create and edit the following file:
+
+`/etc/rsyncd.conf`
+
+```sh
+pid file = /var/run/rsyncd.pid
+lock file = /var/run/rsync.lock
+log file = /var/log/rsync.log
+port = 34873
+max connections = 20
+exclude = lost+found/
+transfer logging = yes
+use chroot = yes
+reverse lookup = no
+
+[gridsnapshots]
+path = /storage/rsync-public/mainnet
+comment = THREEFOLD GRID MAINNET SNAPSHOTS
+read only = true
+timeout = 300
+list = false
+
+[gridsnapshotstest]
+path = /storage/rsync-public/testnet
+comment = THREEFOLD GRID TESTNET SNAPSHOTS
+read only = true
+timeout = 300
+list = false
+
+[gridsnapshotsdev]
+path = /storage/rsync-public/devnet
+comment = THREEFOLD GRID DEVNET SNAPSHOTS
+read only = true
+timeout = 300
+list = false
+```
+
+### Start the Service
+
+Start and enable via systemd:
+
+```sh
+systemctl start rsync
+systemctl enable rsync
+systemctl status rsync
+```
+
+If you're interested about hosting your own instance of the grid to strenghten the ThreeFold ecosystem, make sure to read the next section, [Guardians of the Grid](./tfgrid_guardians.md).
\ No newline at end of file
diff --git a/collections/documentation/developers/grid_deployment/tfgrid_stacks.md b/collections/documentation/developers/grid_deployment/tfgrid_stacks.md
new file mode 100644
index 0000000..7845722
--- /dev/null
+++ b/collections/documentation/developers/grid_deployment/tfgrid_stacks.md
@@ -0,0 +1,32 @@
+ TFGrid Stacks
+Table of Contents
+
+
+- [Introduction](#introduction)
+- [Advantages](#advantages)
+- [Run Your Own Stack](#run-your-own-stack)
+
+
+***
+
+## Introduction
+
+ThreeFold is an open-source project and anyone can run the full stack of the TFGrid in a totally decentralized manner. In practice, this means that anyone can grab a docker compose file shared by ThreeFold of the TFGrid stacks and run an instance of the grid services on their own domain.
+
+This means that you could host your own instance of the ThreeFold Dashboard at `dashboard.yourdomain.com` that would serve your own instance of the complete TFGrid stack! Users could then access the ThreeFold Dashboard via your own domain.
+
+The process is actually very straightforward and we even provide a script to streamline the process.
+
+## Advantages
+
+Setting such instances of the TFGrid ensures resiliency and decentralization of the ThreeFold ecosystem.
+
+As a very concrete example, image that one instance of the Dashboard goes offline, `dashboard.grid.tf`, then users could still access the Dashboard from another instance. The more users of the TFGrid deploy their own instance, the more resilient the grid becomes.
+
+The overall ThreeFold ecosystem becomes more resilient to failures of individual nodes.
+
+## Run Your Own Stack
+
+To set your own instance of the TFGrid, you can download a snapshot of the grid and deploy the TFGrid services with Docker. We even provide scripts to quicken the whole process!
+
+Read more about snapshots in the [next section](./grid_deployment_full_vm.md).
\ No newline at end of file
diff --git a/collections/documentation/developers/internals/internals.md b/collections/documentation/developers/internals/internals.md
new file mode 100644
index 0000000..af1c1f8
--- /dev/null
+++ b/collections/documentation/developers/internals/internals.md
@@ -0,0 +1,19 @@
+ Internals
+
+We present in this section of the developers book a partial list of system components. Content will be added progressively.
+
+ Table of Contents
+
+- [Reliable Message Bus (RMB)](rmb/rmb_toc.md)
+ - [Introduction to RMB](rmb/rmb_intro.md)
+ - [RMB Specs](rmb/rmb_specs.md)
+ - [RMB Peer](rmb/uml/peer.md)
+ - [RMB Relay](rmb/uml/relay.md)
+
+- [ZOS](zos/index.md)
+ - [Manual](./zos/manual/manual.md)
+ - [Workload Types](./zos/manual/workload_types.md)
+ - [Internal Modules](./zos/internals/internals.md)
+ - [Capacity](./zos/internals/capacity.md)
+ - [Performance Monitor Package](./zos/performance/performance.md)
+ - [API](./zos/manual/api.md)
\ No newline at end of file
diff --git a/collections/documentation/developers/internals/rmb/img/layout.png b/collections/documentation/developers/internals/rmb/img/layout.png
new file mode 100644
index 0000000..a6fc981
Binary files /dev/null and b/collections/documentation/developers/internals/rmb/img/layout.png differ
diff --git a/collections/documentation/developers/internals/rmb/img/peer.png b/collections/documentation/developers/internals/rmb/img/peer.png
new file mode 100644
index 0000000..6d6a7d8
Binary files /dev/null and b/collections/documentation/developers/internals/rmb/img/peer.png differ
diff --git a/collections/documentation/developers/internals/rmb/img/relay.png b/collections/documentation/developers/internals/rmb/img/relay.png
new file mode 100644
index 0000000..6b8bf01
Binary files /dev/null and b/collections/documentation/developers/internals/rmb/img/relay.png differ
diff --git a/collections/documentation/developers/internals/rmb/rmb_intro.md b/collections/documentation/developers/internals/rmb/rmb_intro.md
new file mode 100644
index 0000000..bb08f99
--- /dev/null
+++ b/collections/documentation/developers/internals/rmb/rmb_intro.md
@@ -0,0 +1,107 @@
+ Introduction to Reliable Message Bus (RMB)
+
+ Table of Contents
+
+- [What is RMB](#what-is-rmb)
+- [Why](#why)
+- [Specifications](#specifications)
+- [How to Use RMB](#how-to-use-rmb)
+- [Libraries](#libraries)
+ - [Known Libraries](#known-libraries)
+ - [No Known Libraries](#no-known-libraries)
+- [What is rmb-peer](#what-is-rmb-peer)
+- [Download](#download)
+- [Building](#building)
+- [Running tests](#running-tests)
+
+***
+
+## What is RMB
+
+Reliable message bus is a secure communication panel that allows `bots` to communicate together in a `chat` like way. It makes it very easy to host a service or a set of functions to be used by anyone, even if your service is running behind NAT.
+
+Out of the box RMB provides the following:
+
+- Guarantee authenticity of the messages. You are always sure that the received message is from whoever is pretending to be
+- End to End encryption
+- Support for 3rd party hosted relays. Anyone can host a relay and people can use it safely since there is no way messages can be inspected while using e2e. That's similar to `home` servers by `matrix`
+
+![layout](img/layout.png)
+***
+## Why
+
+RMB is developed by ThreefoldTech to create a global network of nodes that are available to host capacity. Each node will act like a single bot where you can ask to host your capacity. This enforced a unique set of requirements:
+
+- Communication needed to be reliable
+ - Minimize and completely eliminate message loss
+ - Reduce downtime
+- Node need to authenticate and authorize calls
+ - Guarantee identity of the other peer so only owners of data can see it
+- Fast request response time
+
+Starting from this we came up with a more detailed requirements:
+
+- User (or rather bots) need their identity maintained by `tfchain` (a blockchain) hence each bot needs an account on tfchain to be able to use `rmb`
+- Then each message then can be signed by the `bot` keys, hence make it easy to verify the identity of the sender of a message. This is done both ways.
+- To support federation (using 3rd party relays) we needed to add e2e encryption to make sure messages that are surfing the public internet can't be sniffed
+- e2e encryption is done by deriving an encryption key from the same identity seed, and share the public key on `tfchain` hence it's available to everyone to use
+***
+## Specifications
+
+For details about protocol itself please check the [specs](./rmb_specs.md).
+***
+## How to Use RMB
+
+There are many ways to use `rmb` because it was built for `bots` and software to communicate. Hence, there is no mobile app for it for example, but instead a set of libraries where you can use to connect to the network, make chitchats with other bots then exit.
+
+Or you can keep the connection forever to answer other bots requests if you are providing a service.
+***
+## Libraries
+
+If there is a library in your preferred language, then you are in luck! Simply follow the library documentations to implement a service bot, or to make requests to other bots.
+
+### Known Libraries
+
+- Golang [rmb-sdk-go](https://github.com/threefoldtech/rmb-sdk-go)
+- Typescript [rmb-sdk-ts](https://github.com/threefoldtech/rmb-sdk-ts)
+***
+### No Known Libraries
+
+If there are no library in your preferred language, here's what you can do:
+
+- Implement a library in your preferred language
+- If it's too much to do all the signing, verification, e2e in your language then use `rmb-peer`
+***
+## What is rmb-peer
+
+think of `rmb-peer` as a gateway that stands between you and the `relay`. `rmb-peer` uses your mnemonics (your identity secret key) to assume your identity and it connects to the relay on your behalf, it maintains the connection forever and takes care of
+
+- reconnecting if connection was lost
+- verifying received messages
+- decrypting received messages
+- sending requests on your behalf, taking care of all crypto heavy lifting.
+
+Then it provide a simple (plain-text) api over `redis`. means to send messages (or handle requests) you just need to be able to push and pop messages from some redis queues. Messages are simple plain text json.
+
+> More details can be found [here](./rmb_specs.md)
+
+***
+## Download
+
+Please check the latest [releases](https://github.com/threefoldtech/rmb-rs/releases) normally you only need the `rmb-peer` binary, unless you want to host your own relay.
+***
+## Building
+
+```bash
+git clone git@github.com:threefoldtech/rmb-rs.git
+cd rmb-rs
+cargo build --release --target=x86_64-unknown-linux-musl
+```
+***
+## Running tests
+
+While inside the repository
+
+```bash
+cargo test
+```
diff --git a/collections/documentation/developers/internals/rmb/rmb_specs.md b/collections/documentation/developers/internals/rmb/rmb_specs.md
new file mode 100644
index 0000000..2d28cbe
--- /dev/null
+++ b/collections/documentation/developers/internals/rmb/rmb_specs.md
@@ -0,0 +1,258 @@
+ RMB Specs
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Overview of the Operation of RMB Relay](#overview-of-the-operation-of-rmb-relay)
+ - [Connections](#connections)
+ - [Peer](#peer)
+ - [Peer implementation](#peer-implementation)
+ - [Message Types](#message-types)
+ - [Output Requests](#output-requests)
+ - [Incoming Response](#incoming-response)
+ - [Incoming Request](#incoming-request)
+ - [Outgoing Response](#outgoing-response)
+- [End2End Encryption](#end2end-encryption)
+- [Rate Limiting](#rate-limiting)
+
+***
+
+# Introduction
+
+RMB is (reliable message bus) is a set of protocols and tools (client and daemon) that aims to abstract inter-process communication between multiple processes running over multiple nodes.
+
+The point behind using RMB is to allow the clients to not know much about the other process, or where it lives (client doesn't know network addresses, or identity). Unlike HTTP(S) or gRPC where the caller must know exact address (or dns-name) and endpoints of the service it's trying to call. Instead RMB requires you to only know about
+
+- Twin ID (numeric ID) of the endpoint as defined by `tfchain`
+- Command (string) is simply the function to call
+- The request "body" which is binary blob that is passed to the command as is
+ - implementation of the command need then to interpret this data as intended (out of scope of rmb)
+
+Twins are stored on tfchain. hence identity of twins is granted not to be spoofed, or phished. When a twin is created he needs to define 2 things:
+
+- RMB Relay
+- His Elliptic Curve public key (we use secp256k1 (K-256) elliptic curve)
+
+> This data is stored on tfchain forever, and only the twin can change it using his secure-key. Hence phishing is impossible. A twin can decide later to change this encryption key or relay.
+
+Once all twins has their data set correctly on the chain. Any 2 twins can communicate with full end-to-end encryption as follows:
+
+- A twin establish a WS connection to his relay of choice
+- A twin create an `envelope` as defined by the protobuf [schema](https://github.com/threefoldtech/rmb-rs/blob/main/proto/types.proto)
+- Twin fill in all envelope information (more about this later)
+- Twin pushes the envelope to the relay
+ - If the destination twin is also using the same relay, message is directly forwarded to this twin
+ - If federation is needed (twin using different relay), message is forwarded to the proper twin.
+
+> NOTE: since a sender twin need to also encrypt the message for the receiver twin, a twin queries the `tf-chain` for the twin information. Usually it caches this data locally for reuse, hence clients need to make sure this data is always up-to-date.
+
+On the relay, the relay checks federation information set on the envelope and then decide to either to forward it internally to one of it's connected clients, or forward it to the destination relay. Hence relays need to be publicly available.
+
+When the relay receive a message that is destined to a `local` connected client, it queue it for delivery. The relay can maintain a queue of messages per twin to a limit. If the twin does not come back online to consume queued messages, the relay will start to drop messages for that specific twin client.
+
+Once a twin come online and connect to its peer, the peer will receive all queued messages. the messages are pushed over the web-socket as they are received. the client then can decide how to handle them (a message can be a request or a response). A message type can be inspected as defined by the schema.
+***
+# Overview of the Operation of RMB Relay
+
+![relay](img/relay.png)
+
+## Connections
+
+By design, there can be only `ONE TWIN` with that specific ID. Hence only has `ONE RELAY` set on tfchain per twin. This force a twin to always use this defined relay if it wishes to open multiple connections to its relay. In other words, a twin once sets up a relay on its public information can only use that relay for all of its connections. If decided to change the relay address, all connections must use the new relay otherwise messages will get lost as they will be delivered to the wrong relay.
+
+In an RPC system, the response of a request must be delivered to the requester. Hence if a twin is maintaining multiple connections to its relay, it need to identify `uniquely` the connection to allow the relay to route back the responses to the right requester. We call this `id` a `session-id`. The `session-id` must be unique per twin.
+
+The relay can maintain **MULTIPLE** connections per peer given that each connection has a unique **SID** (session id). But for each (twin-id, session-id) combo there can be only one connection. if a new connection with the same (twin-id, session-id) is created, the older connection is dropped.
+
+The message received always has the session-id as part of the source address. a reply message then must have destination set back to the source as is, this allows the relay to route the message back correctly without the need to maintain an internal state.
+
+The `rmb-peer` process reserved the `None` sid. It connects with No session id, hence you can only run one `rmb-peer` per `twin` (identity). But the same twin (identity) can make other connection with other rmb clients (for example rmb-sdk-go direct client) to establish more connections with unique session ids.
+
+## Peer
+
+Any language or code that can open `WebSocket` connection to the relay can work as a peer. A peer need to do the following:
+
+- Authenticate with the relay. This is by providing a `JWT` that is signed by the twin key (more on that later)
+- Handle received binary mesasge
+- Send binary messages
+
+Each message is an object of type `Envelope` serialized as with protobuf. Type definition can be found under `proto/types.proto`
+
+## Peer implementation
+
+This project already have a peer implementation that works as local peer gateway. By running this peer instance it allows you to
+run multiple services (and clients) behind that gateway and they appear to the world as a single twin.
+
+- The peer gateway (rmb-peer) starts and connects to realy
+- If requests are received, they are verified, decrypted and pushed to a redis queue that as command specific (from the envelope)
+- A service can then be waiting on this redis queue for new messages
+ - The service can process the command, and push a response back to a specific redis queue for responses.
+- The gateway can then pull ready responses from the responses queue, create a valid envelope, encrypt, and sign and send to destination
+
+![peer](img/peer.png)
+
+### Message Types
+
+Concerning, `rmb-peer` message types, to make it easy for apps to work behind an `rmb-peer`, we use JSON message for communication between the local process and the rmb-peer. the rmb-peer still
+maintains a fully binary communication with the relay.
+
+A request message is defined as follows
+
+#### Output Requests
+
+This is created by a client who wants to request make a request to a remote service
+
+> this message is pushed to `msgbus.system.local` to be picked up by the peer
+
+```rust
+#[derive(Serialize, Deserialize, Clone, Debug)]
+pub struct JsonOutgoingRequest {
+ #[serde(rename = "ver")]
+ pub version: usize,
+ #[serde(rename = "ref")]
+ pub reference: Option,
+ #[serde(rename = "cmd")]
+ pub command: String,
+ #[serde(rename = "exp")]
+ pub expiration: u64,
+ #[serde(rename = "dat")]
+ pub data: String,
+ #[serde(rename = "tag")]
+ pub tags: Option,
+ #[serde(rename = "dst")]
+ pub destinations: Vec,
+ #[serde(rename = "ret")]
+ pub reply_to: String,
+ #[serde(rename = "shm")]
+ pub schema: String,
+ #[serde(rename = "now")]
+ pub timestamp: u64,
+}
+```
+
+#### Incoming Response
+
+A response message is defined as follows this is what is received as a response by a client in response to his outgoing request.
+
+> This response is what is pushed to `$ret` queue defined by the outgoing request, hence the client need to wait on this queue until the response is received or it times out
+
+```rust
+#[derive(Serialize, Deserialize, Clone, Debug)]
+pub struct JsonError {
+ pub code: u32,
+ pub message: String,
+}
+
+#[derive(Serialize, Deserialize, Clone, Debug)]
+pub struct JsonIncomingResponse {
+ #[serde(rename = "ver")]
+ pub version: usize,
+ #[serde(rename = "ref")]
+ pub reference: Option,
+ #[serde(rename = "dat")]
+ pub data: String,
+ #[serde(rename = "src")]
+ pub source: String,
+ #[serde(rename = "shm")]
+ pub schema: Option,
+ #[serde(rename = "now")]
+ pub timestamp: u64,
+ #[serde(rename = "err")]
+ pub error: Option,
+}
+```
+
+#### Incoming Request
+
+An incoming request is a modified version of the request that is received by a service running behind RMB peer
+> this request is received on `msgbus.${request.cmd}` (always prefixed with `msgbus`)
+
+```rust
+#[derive(Serialize, Deserialize, Clone, Debug)]
+pub struct JsonIncomingRequest {
+ #[serde(rename = "ver")]
+ pub version: usize,
+ #[serde(rename = "ref")]
+ pub reference: Option,
+ #[serde(rename = "src")]
+ pub source: String,
+ #[serde(rename = "cmd")]
+ pub command: String,
+ #[serde(rename = "exp")]
+ pub expiration: u64,
+ #[serde(rename = "dat")]
+ pub data: String,
+ #[serde(rename = "tag")]
+ pub tags: Option,
+ #[serde(rename = "ret")]
+ pub reply_to: String,
+ #[serde(rename = "shm")]
+ pub schema: String,
+ #[serde(rename = "now")]
+ pub timestamp: u64,
+}
+```
+
+Services that receive this needs to make sure their responses `destination` to have the same value as the incoming request `source`
+
+#### Outgoing Response
+
+A response message is defined as follows this is what is sent as a response by a service in response to an incoming request.
+
+Your bot (server) need to make sure to set `destination` to the same value as the incoming request `source`
+
+The
+> this response is what is pushed to `msgbus.system.reply`
+
+```rust
+#[derive(Serialize, Deserialize, Clone, Debug)]
+pub struct JsonOutgoingResponse {
+ #[serde(rename = "ver")]
+ pub version: usize,
+ #[serde(rename = "ref")]
+ pub reference: Option,
+ #[serde(rename = "dat")]
+ pub data: String,
+ #[serde(rename = "dst")]
+ pub destination: String,
+ #[serde(rename = "shm")]
+ pub schema: Option,
+ #[serde(rename = "now")]
+ pub timestamp: u64,
+ #[serde(rename = "err")]
+ pub error: Option,
+}
+```
+***
+# End2End Encryption
+
+Relay is totally opaque to the messages. Our implementation of the relay does not poke into messages except for the routing attributes (source, and destinations addresses, and federation information). But since the relay is designed to be hosted by other 3rd parties (hence federation) you should
+not fully trust the relay or whoever is hosting it. Hence e2e was needed
+
+As you already understand e2e is completely up to the peers to implement, and even other implementations of the peers can agree on a completely different encryption algorithm and key sharing algorithm (again, relay does not care). But in our implementation of the e2e (rmb-peer) things goes like this
+
+- Each twin has a `pk` field on tfchain. when rmb-peer start, it generates an `secp256k1` key from the same seed as the user tfchain mnemonics. Note that this will not make the encryption key and the signing key any related, they just are driven from the same seed.
+- On start, if the key is not already set on the twin object, the key is updated.
+- If a peer A is trying to send a message to peer B. but peer B does not has his `pk` set, peer A will send the message in plain-text format (please check the protobuf envelope type for details)
+- If peer B has public key set, peer A will prefer e2e encryption and will does the following:
+- Drive a shared secret point with `ecdh` algorithm, the key is the `sha256` of that point
+- `shared = ecdh(A.sk, B.pk)`
+- create a 12 bytes random nonce
+- encrypt data as `encrypted = aes-gcm.encrypt(shared-key, nonce, plain-data)`
+- create cipher as `cipher nonce + encrypted`
+- fill `envelope.cipher = cipher`
+- on receiving a message peer B does the same in the opposite direction
+- split data and nonce (nonce is always first 12 bytes)
+- derive the same shared key
+- `shared = ecdh(B.sk, A.pk)`
+- `plain-data = aes-gcm.decrypt(shared-key, nonce, encrypted)`
+***
+# Rate Limiting
+
+To avoid abuse of the server, and prevent DoS attacks on the relay, a rate limiter is used to limit the number of clients' requests.\
+It was decided that the rate limiter should only watch websocket connections of users, since all other requests/connections with users consume little resources, and since the relay handles the max number of users inherently.\
+The limiter's configurations are passed as a command line argument `--limit , `. `` represents the number of messages a twin is allowed to send in each time window, `` represents the total size of messages in bytes a twin is allowed to send in each time window.\
+Currently there are two implementations of the rate limiter:
+
+- `NoLimit` which imposes no limits on users.
+- `FixedWindowLimiter` which breaks the timeline into fixed time windows, and allows a twin to send a fixed number of messages, with a fixed total size, in each time window. If a twin exceeded their limits in some time window, their message is dropped, an error message is sent back to the user, the relay dumps a log about this twin, and the user gets to keep their connection with the relay.
diff --git a/collections/documentation/developers/internals/rmb/rmb_toc.md b/collections/documentation/developers/internals/rmb/rmb_toc.md
new file mode 100644
index 0000000..0801571
--- /dev/null
+++ b/collections/documentation/developers/internals/rmb/rmb_toc.md
@@ -0,0 +1,18 @@
+ Reliable Message Bus (RMB)
+
+Reliable message bus is a secure communication panel that allows bots to communicate together in a chat like way. It makes it very easy to host a service or a set of functions to be used by anyone, even if your service is running behind NAT.
+
+Out of the box RMB provides the following:
+
+- Guarantee authenticity of the messages.
+ - You are always sure that the received message is from whoever is pretending to be.
+- End to End encryption.
+- Support for 3rd party hosted relays.
+ - Anyone can host a relay and people can use it safely since there is no way messages can be inspected while using e2e. That's similar to home servers by matrix.
+
+ Table of Contents
+
+- [Introduction to RMB](rmb_intro.md)
+- [RMB Specs](rmb_specs.md)
+- [RMB Peer](uml/peer.md)
+- [RMB Relay](uml/relay.md)
\ No newline at end of file
diff --git a/collections/documentation/developers/internals/rmb/uml/peer.md b/collections/documentation/developers/internals/rmb/uml/peer.md
new file mode 100644
index 0000000..80c195a
--- /dev/null
+++ b/collections/documentation/developers/internals/rmb/uml/peer.md
@@ -0,0 +1,44 @@
+ RMB Peer
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Example](#example)
+
+***
+
+## Introduction
+
+We present an example of RMB peer. Note that the extension for this kind of file is `.wsd`.
+
+## Example
+
+```
+@startuml RMB
+
+participant "Local Process" as ps
+database "Local Redis" as redis
+participant "Rmb Peer" as peer
+
+participant "Rmb Relay" as relay
+note across: Handling Out Request
+peer --> relay: Establish connection
+
+ps -> redis: PUSH message on \n(msgbus.system.local)
+redis -> peer : POP message from \n(msgbus.system.local)
+
+peer -> relay: message pushed over the websocket to the relay
+...
+relay -> peer: received response
+peer -> redis: PUSH over $msg.reply_to queue
+...
+note across: Handling In Request
+relay --> peer: Received a request
+peer -> redis: PUSh request to `msgbus.$cmd`
+redis -> ps: POP new request msg
+ps -> ps: Process message
+ps -> redis: PUSH to (msgbus.system.reply)
+redis -> peer: POP from (msgbus.system.reply)
+peer -> relay: send response message
+@enduml
+```
\ No newline at end of file
diff --git a/collections/documentation/developers/internals/rmb/uml/relay.md b/collections/documentation/developers/internals/rmb/uml/relay.md
new file mode 100644
index 0000000..0618d7b
--- /dev/null
+++ b/collections/documentation/developers/internals/rmb/uml/relay.md
@@ -0,0 +1,40 @@
+ RMB Peer
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Example](#example)
+
+***
+
+## Introduction
+
+We present an example of RMB relay. Note that the extension for this kind of file is `.wsd`.
+
+## Example
+
+
+```
+@startuml RMB
+actor "Peer 1" as peer1
+participant "Relay 1" as relay1
+participant "Relay 2" as relay2
+actor "Peer 2" as peer2
+actor "Peer 3" as peer3
+
+peer1 --> relay1: Establish WS connection
+peer2 --> relay1: Establish WS connection
+peer3 --> relay2: Establish WS connection
+
+peer1 -> relay1: Send message (Envelope)\n(destination "Peer 2")
+relay1 -> peer2: Forward message directly
+
+peer1 -> relay1: Send message (Envelope)\n(destination "Peer 3")
+note right
+"Peer 3" does not live on "Relay 1" hence federation is
+needed
+end note
+relay1 -> relay2: Federation of message for\n Peer 3
+relay2 -> peer3: Forward message directly
+@enduml
+```
\ No newline at end of file
diff --git a/collections/documentation/developers/internals/zos/assets/.keep b/collections/documentation/developers/internals/zos/assets/.keep
new file mode 100644
index 0000000..e69de29
diff --git a/collections/documentation/developers/internals/zos/assets/0-OS v2 architecture.xml b/collections/documentation/developers/internals/zos/assets/0-OS v2 architecture.xml
new file mode 100644
index 0000000..263e7e3
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/assets/0-OS v2 architecture.xml
@@ -0,0 +1 @@
+3Vlbc6JIFP41Pkp10/fHxNzciUkqJpPKvkyhNMoEaQfxll+/BwFFAWN2NTs1WhXgnL75fd85feg0SGu0uI6c8bBjXB00bOQuGuSiYduYKAyXxLJMLULR1DCIfDdrtDF0/XedGVFmnfqunmw1jI0JYn+8beybMNT9eMvmRJGZbzfzTLA969gZ6JKh23eCsvXFd+NhapUMbew32h8M85kxyjwjJ2+cGSZDxzXzgolcNkgrMiZO70aLlg4S8HJc0n5XNd71wiIdxod06LwFL4/3TaGmC7KInu7i29llU6SjzJxgmv3ghs0DGO/cMzAsrDpeZlDwX1OTO5qTFVFn0MCm48XGCXeD5Iqa911w3gAycAn0bCULJ+oP/RhomkY6nwcWnE6VdsywWs9qx3qR2IfxKAADhttJHJk33TKBicASmlAnq/KDYMfkBP4ghMc+AKTBfj7TUewDtWeZY+S7bjLN+TxZVXfs9JM55yBksEVmGro6wQ6tl1UEO8M/GVMvCqYM/GttRjqOltAk8xLC0i7LbYXPN7LCMrMNC5LKpeZkSh6sR96QDTcZ39Xcv0+dv9vu/Mf74/J1aI+uxY1eNnEF9zvgDwCFcc0PXcec08uboxqYajHBSFrbqIgyKpxVoGKjI8ByNls89cT928/OhedO23ejWfuiApZbP5yCxtE3HYWJjndA2igF12ipIN6iThs28VjyLYkaPHz1yeKtYE8/lZLcy/LHOs11ScsUMI4tySlFXAjFOcZlQvjn+Xie6Oi+9zNJ25AjnJ4O0q7d2ERJWoYkatwpwLwaOPDDt9TvOrEDyWeVuu0rINNudcZPwUXr2v9Fo3v75ZWddxhPu+3lmhaQyLRxJGJdR0uvX0lsX+qedwoCcy+2LSxsJgiWikHi+TDClLQwg50LM4oksRE/Cr3wWGB4L+N3Op6b6O0zjI9b8/nz2fMDX949dB9+hJft8eM245UbHj4d40xLl1YxLu0eWYXyqRiHALUI5VyKpADJK6w9fDMLU06EQpRySTH9ar5bkNMcP4Rd+ROMD0Of3l7Nb5+80ffX6K9Ji8jFAYzbJ2NcY+BcVDGuuCDOKRmHLcBimCAlkLApzqvsesoxwhDjjNsUEa6EUuJUnB9WeJLSLvu9syOGY22xnmf3KzOxy3ucnZIlypDFFSGC0qS4kx+RJCEPE8YUgfKGEKnUUTgqEXInNZFPj+8t8vCOWhPTPEPdJq2oBtMK3fVneXXegan8/iSZaVO+F/xVXUzow37uh4Pc2YtyX872rr1u8D9DFTXF1FfLoi43XwX+JD5u7VUpOPZH1l7WpoqiigDFv03tdVgaKB8IPLThL8LHjb7f6LVnHZXCkoJwhoXAinG1RVzVS5GEMtvmhEq4RZQcqYoq0eQJHPxiF8IffxuM3Nsb3X5rVmyfJX7Sd/cS9D2kieZV0CMtEexRxaMP/N8BrzsmqCfCVpYshkS5lrGRhSWFRCgpU5jLLwS+vE0+jweR4yZvq3Mn7g+TmvaIkXIAXWWC9krm44jYA7U4UUlSuWL1r0WOkHQQqkQNMbHy7OQXb/X538Vft38Uz2KwsqB+5wpLQphCxzmLOYyR/MyyuD1EZuZPfBOeKAA8xxVO5Wbeo5whengAqBqS9gbAadCGx83B/8pX+PcJufwH zZbbbqMwEIafBmlvUhmcQ/dyE9JGq22jNj1IexO54IBbYyPbOdCn3zGYAEn2KHWVG7A/hpnh98wID0+y3bUieXojY8q9AMU7D4deEPj9oA83S4qKjC5HFUgUi51RAxbsnTqIHF2zmOqOoZGSG5Z3YSSFoJHpMKKU3HbNVpJ3o+YkoUdgERF+TJ9ZbNKKXg5Qw2eUJWkd2UfuSUZqYwd0SmK5bSE89fBESWmqVbabUG7Fq3V5jHbJZpbN9dOX3djoaPn9ftmrnF39zSv7T1BUmH92fXuHk2/oDqvgHQ309eh29PWxF1SuN4SvnV5eMOQQZJxC1GFiVwsjlRU5QFAda9DVWbyo2qImELx5rYZ5DR5Spq2uzgkqd4rqXArNXkpkpK00SmzeWwb6Qw2UH2HS8g0RyYyJxBoRQ+AmV3BBvfmilUJ+mMGaHxLOauL8rKSyV8602e+gIg1hgto1ETFcn25aYRofJ7wy/VaWkADhMntqAfoE52eYYVJ4wcQGYZzqQhuaVfuLi4s/9I564djmR3ISMVPY0uBEiFKYXxzNsUuAbW3KIjdF3TmG7spCMBkH4MNSGyXf6ERy0AeHQgqwHMN38AOkbWKQDQ4Hze5B5gB60Dp4vE2ZoQvgNtQWBg8wuaFqxcsGS1kcUwFMybWIqS1gtM8QzCCzg+b+TWf4+3aFOUdlRo2ysjkvfdfgRXe7bcYF/uxY2h4V2EHiRlSy99y0ISxcJ57uytclyeZqOnib3r8OJrOiGIXENXK7K1HvCkr8yhXowUE1Kvk/0bZ1hh8loj/sqoiDYxn94ISMw49S8Xi2hcdteY5S7qU7GynxiYIsZ1Ap5fkq2b/8f0rCtvkXKJ+1/qjw9Ac= nZRLc9owEIB/DTPtgRmDSUqP5dGmzUAOZkJuHWGtLSWy1pEFBn59V7aMMcykSS629Gm1u9pXL5xm+1+G5WKBHFRvGPB9L5z1hsPBaDiinyOHmoyDcQ1SI7kXakEkj+Bh4OlWcig6ghZRWZl3YYxaQ2w7jBmDZVcsQdW1mrMUrkAUM3VN15Jb4V9xE7T8DmQqGsuDwJ9krBH2oBCMY3mGwnkvnBpEW6+y/RSUC14Tl9fs7/z+9zJePa/5w7d+eHxdPPVrZT8/cuX0BAPaflp1Bs/s+EP8ye94+phE6/K+NP5KsGNq6+PVG94qMjIRZPU2dasl2BLNC4lRdWwprl5iYxqJhpDx9loD8washCxcXL2SoNoZKHLUhdxUyKKrNGDO71JS/KkGqkdY4Y517YnUKW0woU/Qf4je9Ce/ZFt1SZRsyIYVMnZ2mK1tJEg6g8eF80PzukotkxrMm0Zbjdc2jmDQSqegeU4VFKapEDOX388qxhz0joIWi3demC2jd0rO9Q4P5OWXKj8bqPMGzMQC+NcPekzwPAdVZ9lD064W9lX12UwRGNCysAZfYIqKchHONGqSnCRSqQtU5Cx2WQtnN+1uhTmBPvVrOCmFtBARd6ZKmnYuaDswiaq6WkjOQRMzuNUcXNcEJw9JjDy7mCj/acfBaUbQcAXMwBoXRK9l5Fvv0N2W7YwKv3smzufT2EPm52J60tz2Pi18+zfbdkpVZ2ezPpz/Aw== nZTbjtowEIafJlJ7sVIOsKW3pLRbCbSVoN3V3qxMMsQWju06DoF9+o4Tm5CAtmqvYn/5Z8aHfxwkaXn8pomiK5kDD+IwPwbJlyCOo0k8wY8lp47MwlkHCs1yJ+rBmr2Bg6GjNcuhGgiNlNwwNYSZFAIyM2BEa9kMZTvJh1UVKeAKrDPCr+kTyw11u5iGPX8AVlBfOQrdn5J4sQMVJblsLlCyCJJUS2m6UXlMgdvD8+eyme5f1a4U+2hJl2/Ln9ly9XzXJfv6LyHnLWgQ5r9T/1Y/tpvFcfUkXn89fyoeX2bw4ELCA+G1O68gvudYZE6x6n1hR6kUhjABGoXojxpP1mm22ms8wfJ9oIfKgw1llT1ZlyRsZxoqJUXFti0y0noNiF15w/AGWmd09at366oxy9lhjG4G3tDdQDUfE8480bXIWscIdFNpLykOPzym3z++W7ePv4DXZcYrbK1oTt7fBo7tdZmSI4hwWBkt95BKLjUSIQUq5zvG+QhVimRMFAim/WwjFYI7NHgybygzsEZuSzX4PCCTB9A73rYBZXkOwm5f1iIHa7PwvEKU4cpGLfgX/0bnpsLXCGQJRp8wzmWZOK+ehtOmb+rks2P0sqETB4l7SIpz5r5ZcOD6xU/7tm7/XTyOyeIP rZTBTuMwEIafJkekNCksXBtYWLEcVkUsN2TiaWxhe4LjNC1Pv+PEbppUAq3EKfbn8fzj0T9J8kLvbi2rxQNyUEmW8l2SXydZtlhmS/p4sh/IZXo5gMpKHoJGsJYfEGAaaCs5NJNAh6icrKewRGOgdBPGrMVuGrZBNVWtWQUnYF0ydUr/Su5EeMV5OvI7kJWIyos0nGgWgwNoBOPYHaH8JskLi+iGld4VoHzzYl9+81/ZD/P+/PwOrHiBpsCr+7Mh2c//uXJ4ggXjvjd1NqTeMtWGfiXZhSKRlSDVi8qvnh4ogozRUkvD4auNh5GQ7ngjwjqCRyEb39KQJO13FpoaTSNfe+TQmwyYL7mT1HqqSlrX9kCzUkhDNvpMv54zLrdz1Ko5UTKSP6BbSnPfv1czQ87QvuGfaY63j+CpCMFJMb2d3D561MGub7nTisCClo2z+AYFKrREDBqKXG2kUjPU1KyUpiJwPu4esSZwRibNV52QDtbEvVRHI04Mt2A3qreykJyDIWaxNRy8VdJDhRRGlc3G6AsPLg6DQX8UQA3O7uleyLIMo7SfbrtxMPOrwMTxUOYBsvAzqA6ZR8PTIng+bsfR7M+OfnD5zT8=
\ No newline at end of file
diff --git a/collections/documentation/developers/internals/zos/assets/0-OS-upgrade.png b/collections/documentation/developers/internals/zos/assets/0-OS-upgrade.png
new file mode 100644
index 0000000..f82c801
Binary files /dev/null and b/collections/documentation/developers/internals/zos/assets/0-OS-upgrade.png differ
diff --git a/collections/documentation/developers/internals/zos/assets/0-OS-upgrade.wsd b/collections/documentation/developers/internals/zos/assets/0-OS-upgrade.wsd
new file mode 100644
index 0000000..26ef653
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/assets/0-OS-upgrade.wsd
@@ -0,0 +1,12 @@
+@startuml
+start
+:power on node;
+repeat
+:mount boot flist;
+:copy files to node root;
+:reconfigure services;
+:restart services;
+repeat while (new flist version?) is (yes)
+ -> power off;
+stop
+@enduml
\ No newline at end of file
diff --git a/collections/documentation/developers/internals/zos/assets/0-OS_v2_architecture.png b/collections/documentation/developers/internals/zos/assets/0-OS_v2_architecture.png
new file mode 100644
index 0000000..d4fdd98
Binary files /dev/null and b/collections/documentation/developers/internals/zos/assets/0-OS_v2_architecture.png differ
diff --git a/collections/documentation/developers/internals/zos/assets/Container_module_flow.png b/collections/documentation/developers/internals/zos/assets/Container_module_flow.png
new file mode 100644
index 0000000..95175ae
Binary files /dev/null and b/collections/documentation/developers/internals/zos/assets/Container_module_flow.png differ
diff --git a/collections/documentation/developers/internals/zos/assets/boot_sequence.plantuml b/collections/documentation/developers/internals/zos/assets/boot_sequence.plantuml
new file mode 100644
index 0000000..ac5a663
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/assets/boot_sequence.plantuml
@@ -0,0 +1,50 @@
+@startuml
+
+package "node-ready"{
+ [local-modprobe]
+ [udev-trigger]
+ [redis]
+ [haveged]
+ [cgroup]
+ [redis]
+}
+
+package "boot" {
+ [storaged]
+ [internet]
+ [networkd]
+ [identityd]
+}
+
+package "internal modules"{
+ [flistd]
+ [containerd]
+ [contd]
+ [upgraded]
+ [provisiond]
+}
+
+[local-modprobe]<-- [udev-trigger]
+[udev-trigger] <-- [storaged]
+[udev-trigger] <-- [internet]
+[storaged] <-- [identityd]
+
+[identityd] <- [networkd]
+
+[internet] <-- [networkd]
+[networkd] <-- [containerd]
+[storaged] <-- [containerd]
+
+[containerd] <-- [contd]
+
+[storaged] <-- [flistd]
+[networkd] <-- [flistd]
+
+[flistd] <-- [upgraded]
+[networkd] <-- [upgraded]
+
+[networkd] <-- [provisiond]
+[flistd] <-- [provisiond]
+[contd] <-- [provisiond]
+
+@enduml
diff --git a/collections/documentation/developers/internals/zos/assets/boot_sequence.png b/collections/documentation/developers/internals/zos/assets/boot_sequence.png
new file mode 100644
index 0000000..9a1fc5e
Binary files /dev/null and b/collections/documentation/developers/internals/zos/assets/boot_sequence.png differ
diff --git a/collections/documentation/developers/internals/zos/assets/grid_provisioning.png b/collections/documentation/developers/internals/zos/assets/grid_provisioning.png
new file mode 100644
index 0000000..3cbb45e
Binary files /dev/null and b/collections/documentation/developers/internals/zos/assets/grid_provisioning.png differ
diff --git a/collections/documentation/developers/internals/zos/assets/grid_provisioning.wsd b/collections/documentation/developers/internals/zos/assets/grid_provisioning.wsd
new file mode 100644
index 0000000..48931fc
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/assets/grid_provisioning.wsd
@@ -0,0 +1,37 @@
+@startuml
+title Provisioning of a resource space
+
+autonumber
+actor User as user
+' entity Farmer as farmer
+entity Network as network
+database Blockchain as bc
+boundary Node as node
+collections "Resource space" as rs
+
+== Resource research ==
+user -> network: Send resource request
+activate network
+network -> node: broadcast resource request
+activate node
+deactivate network
+...broadcast to all nodes...
+node -> user: Send offer
+user -> user: inspect offer
+
+== Resource space negotiation ==
+user -> node: accept offer
+user <-> node: key exchange
+user -> bc: money is locked on blockchain
+...
+node -> rs: create resrouce space
+activate rs
+node -> user: notify space is created
+node -> bc: notify he created the space
+user -> rs: make sure it can access the space
+user -> bc: validate can access the space
+bc -> node: money is released to the node
+deactivate node
+== Usage of the space ==
+user -> rs: deploy workload
+@enduml
\ No newline at end of file
diff --git a/collections/documentation/developers/internals/zos/assets/grid_provisioning2.png b/collections/documentation/developers/internals/zos/assets/grid_provisioning2.png
new file mode 100644
index 0000000..2150407
Binary files /dev/null and b/collections/documentation/developers/internals/zos/assets/grid_provisioning2.png differ
diff --git a/collections/documentation/developers/internals/zos/assets/grid_provisioning2.wsd b/collections/documentation/developers/internals/zos/assets/grid_provisioning2.wsd
new file mode 100644
index 0000000..76370a3
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/assets/grid_provisioning2.wsd
@@ -0,0 +1,42 @@
+@startuml
+title Provisioning a workload on the TFGrid
+
+autonumber
+actor "User" as user
+actor "Farmer" as farmer
+database "TF Explorer" as explorer
+database Blockchain as blockchain
+boundary Node as node
+
+== Price definition ==
+farmer -> explorer: Farmer set the price of its Resource units
+== Resource research ==
+activate explorer
+user -> explorer: User look where to deploy the workload
+user <- explorer: Gives detail about the farmer owning the node selected
+== Resource reservation ==
+user -> explorer: write description of the workload
+explorer -> user: return a list of transaction to execute on the blockchain
+== Reservation processing ==
+user -> blockchain: execute transactions
+explorer <-> blockchain: verify transactions are done
+explorer -> explorer: reservation status changed to `deploy`
+== Resource provisioning ==
+node <-> explorer: read description of the workloads
+node -> node: provision workload
+alt provision successfull
+ node -> explorer: write result of the provisining
+ explorer -> blockchain: forward token to the farmer
+ blockchain -> farmer: tokens are available to the farmer
+ user <- explorer: read the connection information to his workload
+else provision error
+ node -> explorer: write result of the provisining
+ explorer -> explorer: cancel reservation
+ node -> node: free up capacity
+ explorer -> blockchain: token refunded to user
+ blockchain <-> user: tokens are available to the user again
+end
+deactivate explorer
+== Resource monitoring ==
+user <-> node: use / monitor workload
+@enduml
\ No newline at end of file
diff --git a/collections/documentation/developers/internals/zos/assets/ipc.plantuml b/collections/documentation/developers/internals/zos/assets/ipc.plantuml
new file mode 100644
index 0000000..20bb31d
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/assets/ipc.plantuml
@@ -0,0 +1,20 @@
+@startuml
+
+== Initialization ==
+Module -> MsgBroker: Announce Module
+MsgBroker -> Module: create bi-directional channel
+
+== Utilisation ==
+loop
+ DSL -> MsgBroker: put RPC message
+ activate MsgBroker
+ Module <- MsgBroker: pull RPC message
+ activate Module
+ Module -> Module: execute method
+ Module -> MsgBroker: put reponse
+ deactivate Module
+ MsgBroker -> DSL : read reponse
+ deactivate MsgBroker
+end
+
+@enduml
\ No newline at end of file
diff --git a/collections/documentation/developers/internals/zos/assets/ipc.png b/collections/documentation/developers/internals/zos/assets/ipc.png
new file mode 100644
index 0000000..9cc940b
Binary files /dev/null and b/collections/documentation/developers/internals/zos/assets/ipc.png differ
diff --git a/collections/documentation/developers/internals/zos/assets/market.png b/collections/documentation/developers/internals/zos/assets/market.png
new file mode 100644
index 0000000..801bdf6
Binary files /dev/null and b/collections/documentation/developers/internals/zos/assets/market.png differ
diff --git a/collections/documentation/developers/internals/zos/assets/market.wsd b/collections/documentation/developers/internals/zos/assets/market.wsd
new file mode 100644
index 0000000..cd49bd7
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/assets/market.wsd
@@ -0,0 +1,22 @@
+@startuml
+actor User as user
+box "To Be Defined" #LightBlue
+ participant Market
+end box
+entity Farmer as farmer
+boundary Node as node
+
+user -> farmer: Request space
+activate farmer
+farmer -> node: reserve space
+activate node
+farmer -> user: confirmation
+deactivate farmer
+...
+note over user, node: communication allows only owner of space
+user -> node: deploy services
+...
+user -> farmer: destroy space
+farmer -> node: delete space
+deactivate node
+@enduml
\ No newline at end of file
diff --git a/collections/documentation/developers/internals/zos/development/README.md b/collections/documentation/developers/internals/zos/development/README.md
new file mode 100644
index 0000000..63b7034
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/development/README.md
@@ -0,0 +1,6 @@
+Development
+===========
+
+* [Quick start](./quickstart.md)
+* [Testing](./testing.md)
+* [Binary packages](./packages.md)
\ No newline at end of file
diff --git a/collections/documentation/developers/internals/zos/development/net.sh b/collections/documentation/developers/internals/zos/development/net.sh
new file mode 100644
index 0000000..ffca7f7
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/development/net.sh
@@ -0,0 +1,30 @@
+#!/bin/bash
+
+# This is the same as the first case at qemu/README.md in a single script
+
+sudo ip link add zos0 type bridge
+sudo ip link set zos0 up
+
+sudo ip addr add 192.168.123.1/24 dev zos0
+md5=$(echo $USER| md5sum )
+ULA=${md5:0:2}:${md5:2:4}:${md5:6:4}
+sudo ip addr add fd${ULA}::1/64 dev zos0
+# you might want to add fe80::1/64
+sudo ip addr add fe80::1/64 dev zos0
+
+sudo iptables -t nat -I POSTROUTING -s 192.168.123.0/24 -j MASQUERADE
+sudo ip6tables -t nat -I POSTROUTING -s fd${ULA}::/64 -j MASQUERADE
+sudo iptables -t filter -I FORWARD --source 192.168.123.0/24 -j ACCEPT
+sudo iptables -t filter -I FORWARD --destination 192.168.123.0/24 -j ACCEPT
+sudo sysctl -w net.ipv4.ip_forward=1
+
+sudo dnsmasq --strict-order \
+ --except-interface=lo \
+ --interface=zos0 \
+ --bind-interfaces \
+ --dhcp-range=192.168.123.20,192.168.123.50 \
+ --dhcp-range=::1000,::1fff,constructor:zos0,ra-stateless,12h \
+ --conf-file="" \
+ --pid-file=/var/run/qemu-dnsmasq-zos0.pid \
+ --dhcp-leasefile=/var/run/qemu-dnsmasq-zos0.leases \
+ --dhcp-no-override
diff --git a/collections/documentation/developers/internals/zos/development/packages.md b/collections/documentation/developers/internals/zos/development/packages.md
new file mode 100644
index 0000000..f7391f6
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/development/packages.md
@@ -0,0 +1,61 @@
+# Adding a new package
+
+Binary packages are added via providing [a build script](../../bins/), then an automated workflow will build/publish an flist with this binary.
+
+For example, to add `rmb` binary, we need to provide a bash script with a `build_rmb` function:
+
+
+```bash
+RMB_VERSION="0.1.2"
+RMB_CHECKSUM="4fefd664f261523b348fc48e9f1c980b"
+RMB_LINK="https://github.com/threefoldtech/rmb-rs/releases/download/v${RMB_VERSION}/rmb"
+
+download_rmb() {
+ echo "download rmb"
+ download_file ${RMB_LINK} ${RMB_CHECKSUM} rmb
+}
+
+prepare_rmb() {
+ echo "[+] prepare rmb"
+ github_name "rmb-${RMB_VERSION}"
+}
+
+install_rmb() {
+ echo "[+] install rmb"
+
+ mkdir -p "${ROOTDIR}/bin"
+
+ cp ${DISTDIR}/rmb ${ROOTDIR}/bin/
+ chmod +x ${ROOTDIR}/bin/*
+}
+
+build_rmb() {
+ pushd "${DISTDIR}"
+
+ download_rmb
+ popd
+
+ prepare_rmb
+ install_rmb
+}
+```
+
+Note that, you can just download a statically build binary instead of building it.
+
+
+The other step is to add it to workflow to be built automatically, in [bins workflow](../../.github/workflows/bins.yaml), add your binary's job:
+
+```yaml
+jobs:
+ containerd:
+ ...
+ ...
+ rmb:
+ uses: ./.github/workflows/bin-package.yaml
+ with:
+ package: rmb
+ secrets:
+ token: ${{ secrets.HUB_JWT }}
+```
+
+Once e.g. a `devnet` release is published, your package will be built then pushed to an flist repository. After that, you can start your local zos node, wait for it to finish downloading, then you should find your binary available.
\ No newline at end of file
diff --git a/collections/documentation/developers/internals/zos/development/quickstart.md b/collections/documentation/developers/internals/zos/development/quickstart.md
new file mode 100644
index 0000000..0563b61
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/development/quickstart.md
@@ -0,0 +1,70 @@
+# Quick start
+
+- [Quick start](#quick-start)
+ - [Starting a local zos node](#starting-a-local-zos-node)
+ - [Accessing node](#accessing-node)
+ - [Development](#development)
+
+## Starting a local zos node
+
+* Make sure `qemu` and `dnsmasq` are installed
+* [Create a farm](../manual/manual.md#creating-a-farm)
+* [Download a zos image](https://bootstrap.grid.tf/kernel/zero-os-development-zos-v3-generic-7e587e499a.efi)
+* Make sure `zos0` bridge is allowed by qemu, you can add `allow zos0` in `/etc/qemu/bridge.conf` (create the file if it's not there)
+* Setup the network using this script [this script](./net.sh)
+
+Then, inside zos repository
+
+```
+make -C cmds
+cd qemu
+mv ./zos.efi
+sudo ./vm.sh -n myzos-01 -c "farmer_id= printk.devmsg=on runmode=dev"
+```
+
+You should see the qemu console and boot logs, wait for awhile and you can [browse farms](https://dashboard.dev.grid.tf/explorer/farms) to see your node is added/detected automatically.
+
+To stop the machine you can do `Control + a` then `x`.
+
+You can read more about setting up a qemu development environment and more network options [here](../../qemu/README.md).
+
+## Accessing node
+
+After booting up, the node should start downloading external packages, this would take some time depending on your internet connection.
+
+See [how to ssh into it.](../../qemu/README.md#to-ssh-into-the-machine)
+
+How to get the node IP?
+Given the network script `dhcp-range`, it usually would be one of `192.168.123.43`, `192.168.123.44` or `192.168.123.45`.
+
+Or you can simply install `arp-scan` then do something like:
+
+```
+✗ sudo arp-scan --interface=zos0 --localnet
+Interface: zos0, type: EN10MB, MAC: de:26:45:e6:87:95, IPv4: 192.168.123.1
+Starting arp-scan 1.9.7 with 256 hosts (https://github.com/royhills/arp-scan)
+192.168.123.44 54:43:83:1f:eb:81 (Unknown)
+```
+
+Now we know for sure it's `192.168.123.44`.
+
+To check logs and see if the downloading of packages is still in progress, you can simply do:
+
+```
+zinit log
+```
+
+## Development
+
+While the overlay will enable your to boot with the binaries that's been built locally, sometimes you'll need to test the changes of certain modules without restarting the node (or intending to do so for e.g. testing a migration).
+
+For example if we changed anything related to `noded`, we can do the following:
+
+Inside zos repository:
+
+* Build binaries locally
+ * `make -C cmds`
+* Copy the binary inside the machine
+ * `scp bin/zos root@192.168.123.44:/bin/noded`
+* SSH into the machine then use `zinit` to restart it:
+ * `zinit stop noded && zinit start noded`
diff --git a/collections/documentation/developers/internals/zos/development/testing.md b/collections/documentation/developers/internals/zos/development/testing.md
new file mode 100644
index 0000000..8f023ec
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/development/testing.md
@@ -0,0 +1,157 @@
+# Testing
+
+Beside unit testing, you might want to test your change in an integrated environment, the following are two options to do it.
+
+- [Testing](#testing)
+ - [Using grid/node client](#using-gridnode-client)
+ - [Using a test app](#using-a-test-app)
+ - [An example to talk to container and qsfs modules](#an-example-to-talk-to-container-and-qsfs-modules)
+ - [An example of directly using zinit package](#an-example-of-directly-using-zinit-package)
+
+
+## Using grid/node client
+
+You can simply use any grid client to deploy a workload of any type, you should specify your node's twin ID (and make sure you are on the correct network).
+
+Inside the node, you can do `noded -id` and `noded -net` to get your current node ID and network. Also, [you can check your farm](https://dashboard.dev.grid.tf/explorer/farms) and get node information from there.
+
+Another option is the golang [node client](../manual/manual.md#interaction).
+
+While deploying on your local node, logs with `zinit log` would be helpful to see any possible errors and to debug your code.
+
+## Using a test app
+
+If you need to test a specific module or functionality, you can create a simple test app inside e.g. [tools directory](../../tools/).
+
+Inside this simple test app, you can import any module or talk to another one using [zbus](../internals/internals.md#ipc).
+
+### An example to talk to container and qsfs modules
+
+
+```go
+// tools/del/main.go
+
+package main
+
+import (
+ "context"
+ "flag"
+ "strings"
+ "time"
+
+ "github.com/rs/zerolog"
+ "github.com/rs/zerolog/log"
+
+ "github.com/threefoldtech/zbus"
+ "github.com/threefoldtech/zos/pkg"
+ "github.com/threefoldtech/zos/pkg/stubs"
+)
+
+func main() {
+ zerolog.SetGlobalLevel(zerolog.DebugLevel)
+
+ zbus, err := zbus.NewRedisClient("unix:///var/run/redis.sock")
+ if err != nil {
+ log.Err(err).Msg("cannot init zbus client")
+ return
+ }
+
+ var workloadType, workloadID string
+
+ flag.StringVar(&workloadType, "type", "", "workload type (qsfs or container)")
+ flag.StringVar(&workloadID, "id", "", "workload ID")
+
+ flag.Parse()
+
+ if workloadType == "" || workloadID == "" {
+ log.Error().Msg("you need to provide both type and id")
+ return
+ }
+
+ ctx, cancel := context.WithTimeout(context.Background(), 10*time.Second)
+ defer cancel()
+
+ if workloadType == "qsfs" {
+ qsfsd := stubs.NewQSFSDStub(zbus)
+ err := qsfsd.SignalDelete(ctx, workloadID)
+ if err != nil {
+ log.Err(err).Msg("cannot delete qsfs workload")
+ }
+ } else if workloadType == "container" {
+ args := strings.Split(workloadID, ":")
+ if len(args) != 2 {
+ log.Error().Msg("container id must contain namespace, e.g. qsfs:wl129")
+ }
+
+ containerd := stubs.NewContainerModuleStub(zbus)
+ err := containerd.SignalDelete(ctx, args[0], pkg.ContainerID(args[1]))
+ if err != nil {
+ log.Err(err).Msg("cannot delete container workload")
+ }
+ }
+
+}
+```
+
+Then we can simply build, upload and execute this in our node:
+
+```
+cd tools/del
+go build
+scp del root@192.168.123.44:/root/del
+```
+
+Then ssh into `192.168.123.44` and simply execute your test app:
+
+```
+./del
+```
+
+### An example of directly using zinit package
+
+```go
+// tools/zinit_test
+package main
+
+import (
+ "encoding/json"
+ "fmt"
+ "regexp"
+
+ "github.com/rs/zerolog"
+ "github.com/rs/zerolog/log"
+
+ "github.com/threefoldtech/zos/pkg/zinit"
+)
+
+func main() {
+ zerolog.SetGlobalLevel(zerolog.DebugLevel)
+ z := zinit.New("/var/run/zinit.sock")
+
+ regex := fmt.Sprintf(`^ip netns exec %s %s`, "ndmz", "/sbin/udhcpc")
+ _, err := regexp.Compile(regex)
+ if err != nil {
+ log.Err(err).Msgf("cannot compile %s", regex)
+ return
+ }
+
+ // try match
+ matched, err := z.Matches(zinit.WithExecRegex(regex))
+ if err != nil {
+ log.Err(err).Msg("cannot filter services")
+ }
+
+ matchedStr, err := json.Marshal(matched)
+ if err != nil {
+ log.Err(err).Msg("cannot convert matched map to json")
+ }
+
+ log.Debug().Str("matched", string(matchedStr)).Msg("matched services")
+
+ // // try destroy
+ // err = z.Destroy(10*time.Second, matched...)
+ // if err != nil {
+ // log.Err(err).Msg("cannot destroy matched services")
+ // }
+}
+```
\ No newline at end of file
diff --git a/collections/documentation/developers/internals/zos/faq/readme.md b/collections/documentation/developers/internals/zos/faq/readme.md
new file mode 100644
index 0000000..f241686
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/faq/readme.md
@@ -0,0 +1,6 @@
+# FAQ
+
+This section consolidated all the common question we get about how 0-OS work and how to operate it.
+
+- **Q**: What is the preferred configuration for my raid controller when running 0-OS ?
+ **A**: 0-OS goal is to expose raw capacity. So it is best to always try to give him access to the most raw access to the disks. In case of raid controllers, the best is to try to set it up in [JBOD](https://en.wikipedia.org/wiki/Non-RAID_drive_architectures#JBOD) mode if available.
diff --git a/collections/documentation/developers/internals/zos/internals/boot.md b/collections/documentation/developers/internals/zos/internals/boot.md
new file mode 100644
index 0000000..bb98c8a
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/boot.md
@@ -0,0 +1,11 @@
+# Services Boot Sequence
+
+Here is dependency graph of all the services started by 0-OS:
+
+![boot sequence](../assets/boot_sequence.png)
+
+## Pseudo boot steps
+
+both `node-ready` and `boot` are not actual services, but instead they are there to define a `boot stage`. for example once `node-ready` service is (ready) it means all crucial system services defined by 0-initramfs are now running.
+
+`boot` service is similar, but guarantees that some 0-OS services are running (for example `storaged`), before starting other services like `flistd` which requires `storaged`
diff --git a/collections/documentation/developers/internals/zos/internals/capacity.md b/collections/documentation/developers/internals/zos/internals/capacity.md
new file mode 100644
index 0000000..04d1137
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/capacity.md
@@ -0,0 +1,89 @@
+Capacity
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [System reserved capacity](#system-reserved-capacity)
+ - [Reserved Memory](#reserved-memory)
+ - [Reserved Storage](#reserved-storage)
+- [User Capacity](#user-capacity)
+ - [Memory](#memory)
+ - [Storage](#storage)
+
+***
+
+## Introduction
+
+This document describes how ZOS does the following tasks:
+
+- Reserved system resources
+ - Memory
+ - Storage
+- Calculation of free usable capacity for user workloads
+
+## System reserved capacity
+
+ZOS always reserve some amount of the available physical resources to its own operation. The system tries to be as protective
+as possible of it's critical services to make sure that the node is always reachable and usable even if it's under heavy load
+
+ZOS make sure it reserves Memory and Storage (but not CPU) as per the following:
+
+### Reserved Memory
+
+ZOS reserve 10% of the available system memory for basic services AND operation overhead. The operation overhead can happen as a side effect of running user workloads. For example, a user network while in theory does not consume any memory, in matter of fact it also consume some memory (kernel buffers, etc...). Same for a VM. A user VM can be assigned say 5G but the process that running the VM can/will take few extra megabytes to operate.
+
+This is why we decided to play on the safe side, and reserve 10% of total system memory to the system overhead, with a **MIN** reserved memory of 2GB
+
+```python
+reserved = min(total_in_gb * 0.1, 2G)
+```
+
+### Reserved Storage
+
+While ZOS does not require installation, but it needs to download and store many things to operate correctly. This include the following:
+
+- Node identity. Information about the node id and keys
+- The system binaries, those what include all zos to join the grid and operate as expected
+- Workload flists. Those are the flists of the user workloads. Those are downloaded on demand so they don't always exist.
+- State information. Tracking information maintained by ZOS to track the state of workloads, owner-ship, and more.
+
+This is why the system on first start allocates and reserve a part of the available SSD storage and is called `zos-cache`. Initially is `5G` (was 100G in older version) but because the `dynamic` nature of the cache we can't fix it at `5G`
+
+The required space to be reserved by the system can dramatically change based on the amount of workloads running on the system. For example if many users are running many different VMs, the system will need to download (and cache) different VM images, hence requiring more cache.
+
+This is why the system periodically checks the reserved storage and then dynamically expand or shrink to a more suitable value in increments of 5G. The expansion happens around the 20% of current cache size, and shrinking if went below 20%.
+
+## User Capacity
+
+All workloads requires some sort of a resource(s) to run and that is actually what the user hae to pay for. Any workload can consume resources in one of the following criteria:
+
+- CU (compute unit in vCPU)
+- MU (memory unit in bytes)
+- NU (network unit in bytes)
+- SU (ssd storage in bytes)
+- HU (hdd storage in bytes)
+
+A workloads, based on the type can consume one or more of those resource types. Some workloads will have a well known "size" on creation, others might be dynamic and won't be know until later.
+
+For example, a disk workload SU consumption will be know ahead. Unlike the NU used by a network which will only be known after usage over a certain period of time.
+
+A single deployment can has multiple workloads each requires a certain amount of one or more capacity types (listed above). ZOS then for each workloads type compute the amount of resources needed per workload, and then check if it can provide this amount of capacity.
+
+> This means that a deployment that define 2 VMs can partially succeed to deploy one of the VMs but not the other one if the amount of resources it requested are higher than what the node can provide
+
+### Memory
+
+How the system decide if there are enough memory to run a certain workload that demands MU resources goes as follows:
+
+- compute the "theoretically used" memory by all user workloads excluding `self`. This is basically the sum of all consumed MU units of all active workloads (as defined by their corresponding deployments, not as per actually used in the system).
+- The theoretically used memory is topped with the system reserved memory.
+- The the system checks actually used memory on the system this is done simply by doing `actual_used = memory.total - memory.available`
+- The system now can simply `assume` an accurate used memory by doing `used = max(actual_used, theoretically_used)`
+- Then `available = total - used`
+- Then simply checks that `available` memory is enough to hold requested workload memory!
+
+### Storage
+
+Storage is much simpler to allocate than memory. It's completely left to the storage subsystem to find out if it can fit the requested storage on the available physical disks or not, if not possible the workloads is marked as error.
+
+Storage tries to find the requested space based on type (SU or HU), then find the optimal way to fit that on the available disks, or spin up a new one if needed.
diff --git a/collections/documentation/developers/internals/zos/internals/compatibility/readme.md b/collections/documentation/developers/internals/zos/internals/compatibility/readme.md
new file mode 100644
index 0000000..1eb4394
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/compatibility/readme.md
@@ -0,0 +1,14 @@
+# Compatibility list
+
+This document track all the hardware that have been tested, the issues encountered and possible workarounds.
+
+**Legend**
+✅ : fully supported
+⚠️ : supported with some tweaking
+🛑 : not supported
+
+
+| vendor | Hardware | Support | Issues | workaround |
+| --- | --- | --- | --- | --- |
+| Supermicro | SYS-5038ML-H8TRF | ✅ | | |
+| Gigabyte Technology Co | AB350N-Gaming WIFI | ✅ | | |
diff --git a/collections/documentation/developers/internals/zos/internals/container/readme.md b/collections/documentation/developers/internals/zos/internals/container/readme.md
new file mode 100644
index 0000000..6d5d2e2
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/container/readme.md
@@ -0,0 +1,106 @@
+Container Module
+
+ Table of Contents
+
+- [ZBus](#zbus)
+- [Home Directory](#home-directory)
+- [Introduction](#introduction)
+ - [zinit unit](#zinit-unit)
+- [Interface](#interface)
+
+***
+
+## ZBus
+
+Storage module is available on zbus over the following channel
+
+| module | object | version |
+|--------|--------|---------|
+| container|[container](#interface)| 0.0.1|
+
+## Home Directory
+
+contd keeps some data in the following locations
+| directory | path|
+|----|---|
+| root| `/var/cache/modules/containerd`|
+
+## Introduction
+
+The container module, is a proxy to [containerd](https://github.com/containerd/containerd). The proxy provides integration with zbus.
+
+The implementation is the moment is straight forward, which includes preparing the OCI spec for the container, the tenant containerd namespace,
+setting up proper capabilities, and finally creating the container instance on `containerd`.
+
+The module is fully stateless, all container information is queried during runtime from `containerd`.
+
+### zinit unit
+
+`contd` must run after containerd is running, and the node boot process is complete. Since it doesn't keep state, no dependency on `stroaged` is needed
+
+```yaml
+exec: contd -broker unix:///var/run/redis.sock -root /var/cache/modules/containerd
+after:
+ - containerd
+ - boot
+```
+
+## Interface
+
+```go
+package pkg
+
+// ContainerID type
+type ContainerID string
+
+// NetworkInfo defines a network configuration for a container
+type NetworkInfo struct {
+ // Currently a container can only join one (and only one)
+ // network namespace that has to be pre defined on the node
+ // for the container tenant
+
+ // Containers don't need to know about anything about bridges,
+ // IPs, wireguards since this is all is only known by the network
+ // resource which is out of the scope of this module
+ Namespace string
+}
+
+// MountInfo defines a mount point
+type MountInfo struct {
+ Source string // source of the mount point on the host
+ Target string // target of mount inside the container
+ Type string // mount type
+ Options []string // mount options
+}
+
+//Container creation info
+type Container struct {
+ // Name of container
+ Name string
+ // path to the rootfs of the container
+ RootFS string
+ // Env env variables to container in format {'KEY=VALUE', 'KEY2=VALUE2'}
+ Env []string
+ // Network network info for container
+ Network NetworkInfo
+ // Mounts extra mounts for container
+ Mounts []MountInfo
+ // Entrypoint the process to start inside the container
+ Entrypoint string
+ // Interactivity enable Core X as PID 1 on the container
+ Interactive bool
+}
+
+// ContainerModule defines rpc interface to containerd
+type ContainerModule interface {
+ // Run creates and starts a container on the node. It also auto
+ // starts command defined by `entrypoint` inside the container
+ // ns: tenant namespace
+ // data: Container info
+ Run(ns string, data Container) (ContainerID, error)
+
+ // Inspect, return information about the container, given its container id
+ Inspect(ns string, id ContainerID) (Container, error)
+ Delete(ns string, id ContainerID) error
+}
+```
diff --git a/collections/documentation/developers/internals/zos/internals/flist/readme.md b/collections/documentation/developers/internals/zos/internals/flist/readme.md
new file mode 100644
index 0000000..46f1076
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/flist/readme.md
@@ -0,0 +1,74 @@
+Flist Module
+
+ Table of Contents
+
+- [Zbus](#zbus)
+- [Home Directory](#home-directory)
+- [Introduction](#introduction)
+- [Public interface ](#public-interface-)
+- [zinit unit](#zinit-unit)
+
+***
+
+## Zbus
+
+Flist module is available on zbus over the following channel:
+
+| module | object | version |
+|--------|--------|---------|
+|flist |[flist](#public-interface)| 0.0.1
+
+## Home Directory
+flist keeps some data in the following locations:
+| directory | path|
+|----|---|
+| root| `/var/cache/modules/containerd`|
+
+## Introduction
+
+This module is responsible to "mount an flist" in the filesystem of the node. The mounted directory contains all the files required by containers or (in the future) VMs.
+
+The flist module interface is very simple. It does not expose any way to choose where to mount the flist or have any reference to containers or VM. The only functionality is to mount a given flist and receive the location where it is mounted. It is up to the above layer to do something useful with this information.
+
+The flist module itself doesn't contain the logic to understand the flist format or to run the fuse filesystem. It is just a wrapper that manages [0-fs](https://github.com/threefoldtech/0-fs) processes.
+
+Its only job is to download the flist, prepare the isolation of all the data and then start 0-fs with the proper arguments.
+
+## Public interface [![GoDoc](https://godoc.org/github.com/threefoldtech/zos/pkg/flist?status.svg)](https://godoc.org/github.com/threefoldtech/zos/pkg/flist)
+
+```go
+
+//Flister is the interface for the flist module
+type Flister interface {
+ // Mount mounts an flist located at url using the 0-db located at storage
+ // in a RO mode. note that there is no way u can unmount a ro flist because
+ // it can be shared by many users, it's then up to system to decide if the
+ // mount is not needed anymore and clean it up
+ Mount(name, url string, opt MountOptions) (path string, err error)
+
+ // UpdateMountSize change the mount size
+ UpdateMountSize(name string, limit gridtypes.Unit) (path string, err error)
+
+ // Umount a RW mount. this only unmounts the RW layer and remove the assigned
+ // volume.
+ Unmount(name string) error
+
+ // HashFromRootPath returns flist hash from a running g8ufs mounted with NamedMount
+ HashFromRootPath(name string) (string, error)
+
+ // FlistHash returns md5 of flist if available (requesting the hub)
+ FlistHash(url string) (string, error)
+
+ Exists(name string) (bool, error)
+}
+
+```
+
+## zinit unit
+
+The zinit unit file of the module specifies the command line, test command, and the order in which the services need to be booted.
+
+Flist module depends on the storage and network pkg.
+This is because it needs connectivity to download flist and data and it needs storage to be able to cache the data once downloaded.
+
+Flist doesn't do anything special on the system except creating a bunch of directories it will use during its lifetime.
diff --git a/collections/documentation/developers/internals/zos/internals/gateway/readme.md b/collections/documentation/developers/internals/zos/internals/gateway/readme.md
new file mode 100644
index 0000000..45f5035
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/gateway/readme.md
@@ -0,0 +1,121 @@
+# Gateway Module
+
+## ZBus
+
+Gateway module is available on zbus over the following channel
+
+| module | object | version |
+| ------- | --------------------- | ------- |
+| gateway | [gateway](#interface) | 0.0.1 |
+
+## Home Directory
+
+gateway keeps some data in the following locations
+| directory | path |
+| --------- | ---------------------------- |
+| root | `/var/cache/modules/gateway` |
+
+The directory `/var/cache/modules/gateway/proxy` contains the route information used by traefik to forward traffic.
+## Introduction
+
+The gateway modules is used to register traefik routes and services to act as a reverse proxy. It's the backend supporting two kinds of workloads: `gateway-fqdn-proxy` and `gateway-name-proxy`.
+
+For the FQDN type, it receives the domain and a list of backends in the form `http://ip:port` or `https://ip:port` and registers a route for this domain forwarding traffic to these backends. It's a requirement that the domain resolves to the gateway public ip. The `tls_passthrough` parameter determines whether the tls termination happens on the gateway or in the backends. When it's true, the backends must be in the form `https://ip:port`, and the backends must be https-enabled servers.
+
+The name type is the same as the FQDN type except that the `name` parameter is added as a prefix to the gatweay domain to determine the fqdn. It's forbidden to use a FQDN type workload to reserve a domain managed by the gateway.
+
+The fqdn type is enabled only if there's a public config on the node. The name type works only if a domain exists in the public config. To make a full-fledged gateway node, these DNS records are required:
+```
+gatwaydomain.com A ip.of.the.gateway
+*.gatewaydomain.com CNAME gatewaydomain.com
+__acme-challenge.gatewaydomain.com NS gatdwaydomain.com
+```
+
+### zinit unit
+
+```yaml
+exec: gateway --broker unix:///var/run/redis.sock --root /var/cache/modules/gateway
+after:
+ - boot
+```
+## Implementation details
+
+Traefik is used as the reverse proxy forwarding traffic to upstream servers. All worklaods deployed on the node is associated with a domain that resolves to the node IP. In the name workload case, it's a subdomain of the gateway main domain. In the FQDN case, the user must create a DNS A record pointing it to the node IP. The node by default redirects all http traffic to https.
+
+When an https request reaches the node, it looks at the domain and determines the correct service that should handle the request. The services defintions are in `/var/cache/modules/gateway/proxy/` and is hot-reloaded by traefik every time a service is added/removed to/from it. Zos currently supports enabling `tls_passthrough` in which case the https request is passed as is to the backend (at the TCP level). The default is `tls_passthrough` is false which means the node terminates the TLS traffic and then forwards the request as http to the backend.
+Example of a FQDN service definition with tls_passthrough enabled:
+```yaml
+tcp:
+ routers:
+ 37-2039-testname-route:
+ rule: HostSNI(`remote.omar.grid.tf`)
+ service: 37-2039-testname
+ tls:
+ passthrough: "true"
+ services:
+ 37-2039-testname:
+ loadbalancer:
+ servers:
+ - address: 137.184.106.152:443
+```
+Example of a "name" service definition with tls_passthrough disabled:
+```yaml
+http:
+ routers:
+ 37-1976-workloadname-route:
+ rule: Host(`workloadname.gent01.dev.grid.tf`)
+ service: 40-1976-workloadname
+ tls:
+ certResolver: dnsresolver
+ domains:
+ - sans:
+ - '*.gent01.dev.grid.tf'
+ services:
+ 40-1976-workloadname:
+ loadbalancer:
+ servers:
+ - url: http://[backendip]:9000
+```
+
+The `certResolver` option has two valid values, `resolver` and `dnsresolver`. The `resolver` is an http resolver and is used in FQDN services with `tls_passthrough` disabled. It uses the http challenge to generate a single-domain certificate. The `dnsresolver` is used for name services with `tls_passthrough` disabled. The `dnsresolver` is responsible for generating a wildcard certificate to be used for all subdomains of the gateway domain. Its flow is described below.
+
+The CNAME record is used to make all subdomains (reserved or not) resolve to the ip of the gateway. Generating a wildcard certificate requires adding a TXT record at `__acme-challenge.gatewaydomain.com`. The NS record is used to delegate this specific subdomain to the node. So if someone did `dig TXT __acme-challenge.gatewaydomain.com`, the query is served by the node, not the DNS provider used for the gateway domain.
+
+Traefik has, as a config parameter, multiple dns [providers](https://doc.traefik.io/traefik/https/acme/#providers) to communicate with when it wants to add the required TXT record. For non-supported providers, a bash script can be provided to do the record generation and clean up (i.e. External program). The bash [script](https://github.com/threefoldtech/zos/blob/main/pkg/gateway/static/cert.sh) starts dnsmasq managing a dns zone for the `__acme-challenge` subdomain with the given TXT record. It then kills the dnsmasq process and removes the config file during cleanup.
+## Interface
+
+```go
+type Backend string
+
+// GatewayFQDNProxy definition. this will proxy name. to backends
+type GatewayFQDNProxy struct {
+ // FQDN the fully qualified domain name to use (cannot be present with Name)
+ FQDN string `json:"fqdn"`
+
+ // Passthroug whether to pass tls traffic or not
+ TLSPassthrough bool `json:"tls_passthrough"`
+
+ // Backends are list of backend ips
+ Backends []Backend `json:"backends"`
+}
+
+
+// GatewayNameProxy definition. this will proxy name. to backends
+type GatewayNameProxy struct {
+ // Name the fully qualified domain name to use (cannot be present with Name)
+ Name string `json:"name"`
+
+ // Passthroug whether to pass tls traffic or not
+ TLSPassthrough bool `json:"tls_passthrough"`
+
+ // Backends are list of backend ips
+ Backends []Backend `json:"backends"`
+}
+
+type Gateway interface {
+ SetNamedProxy(wlID string, prefix string, backends []string, TLSPassthrough bool) (string, error)
+ SetFQDNProxy(wlID string, fqdn string, backends []string, TLSPassthrough bool) error
+ DeleteNamedProxy(wlID string) error
+ Metrics() (GatewayMetrics, error)
+}
+```
diff --git a/collections/documentation/developers/internals/zos/internals/history/readme.md b/collections/documentation/developers/internals/zos/internals/history/readme.md
new file mode 100644
index 0000000..e4c023a
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/history/readme.md
@@ -0,0 +1,99 @@
+# 0-OS, a bit of history and introduction to Version 2
+
+## Once upon a time
+----
+A few years ago, we were trying to come up with some solutions to the problem of self-healing IT.
+We boldly started that : the current model of cloud computing in huge data-centers is not going to be able to scale to fit the demand in IT capacity.
+
+The approach we took to solve this problem was to enable localized compute and storage units at the edge of the network, close to where it is needed.
+That basically meant that if we were to deploy physical hardware to the edges, nearby the users, we would have to allow information providers to deploy their solutions on that edge network and hardware. That means also sharing hardware resources between users, where we would have to make damn sure noone can peek around in things that are not his.
+
+When we talk about sharing capacity in a secure environment, virtualization comes to mind. It's not a new technology and it has been around for quite some time. This solution comes with a cost though. Virtual machines, emulating a full hardware platform on real hardware is costly in terms of used resources, and eat away at the already scarce resources we want to provide for our users.
+
+Containerizing technologies were starting to get some hype at the time. Containers provide for basically the same level of isolation as Full Virtualisation, but are a lot less expensive in terms of resource utilization.
+
+With that in mind, we started designing the first version of 0-OS. The required features were:
+
+- be able to be fully in control of the hardware
+- give the possibility to different users to share the same hardware
+- deploy this capacity at the edge, close to where it is needed
+- the System needs to self-heal. Because of their location and sheer scale, manual maintenance was not an option. Self-healing is a broad topic, and will require a lot of experience and fine-tuning, but it was meant to culminate at some point so that most of the actions that sysadmins execute, would be automated.
+- Have an a small as possible attack surface, as well for remote types of attack, as well as protecting users from each-other
+
+The result of that thought process resulted in 0-OS v1. A linux kernel with the minimal components on top that allows to provide for these features.
+
+In the first incantation of 0-OS, the core framework was a single big binary that got started as the first process of the system (PID 1). All the managment features were exposed through an API that was only accessible locally.
+
+The idea was to have an orchestration system running on top that was going to be responsible to deploy Virtual Machines and Containers on the system using that API.
+
+This API exposes 3 main primitives:
+
+- networking: zerotier, vlan, macvlan, bridge, openvswitch...
+- storage: plain disk, 0-db, ...
+- compute: VM, containers
+
+That was all great and it allowed us to learn a lot. But some limitations started to appear. Here is a non exhaustive list of the limitations we had to face after a couple of years of utilization:
+
+- Difficulty to push new versions and fixes on the nodes. The fact that 0-OS was a single process running as PID 1, forced us to completely reboot the node every time we wanted to push an update.
+- The API, while powerful, still required to have some logic on top to actually deploy usable solutions.
+- We noticed that some features we implemented were never or extremely rarely used. This was just increasing the possible attack surface for no real benefits.
+- The main networking solution we choose at the time, zerotier, was not scaling as well as we hoped for.
+- We wrote a lot of code ourselves, instead of relying on already existing open source libraries that would have made that task a lot easier, but also, these libraries were a lot more mature and have had a lot more exposure for ironing out possible bugs and vulnerabilities than we could have created and tested ourselves with the little resources we have at hand.
+
+## Now what ?
+With the knowledge and lessons gathered during these first years of usage, we
+concluded that trying to fix the already existing codebase would be cumbersome
+and we also wanted to avoid any technical debt that could haunt us for years
+after. So we decided for a complete rewrite of that stack, taking a new and
+fully modular approach, where every component could be easily replaced and
+upgraded without the need for a reboot.
+
+Hence Version 2 saw the light of day.
+
+Instead of trial and error, and muddling along trying to fit new features in
+that big monolithic codebase, we wanted to be sure that the components were
+reduced to a more manageable size, having a clearly cut Domain Separation.
+
+Instead of creating solutions waiting for a problem, we started looking at things the other way around. Which is logical, as by now, we learned what the real puzzles to solve were, albeit sometimes by painful experience.
+
+## Tadaa!
+----
+The [first commit](https://github.com/threefoldtech/zosv2/commit/7b783c888673d1e9bc400e4abbb17272e995f5a4) of the v2 repository took place the 11 of February 2019.
+We are now 6 months in, and about to bake the first release of 0-OS v2.
+Clocking in at almost 27KLoc, it was a very busy half-year. (admitted, there are the spec and docs too in that count ;-) )
+
+Let's go over the main design decisions that were made and explain briefly each component.
+
+While this is just an introduction, we'll add more articles digging deeper in the technicalities and approaches of each component.
+
+## Solutions to puzzles (there are no problems)
+----
+**UPDATES**
+
+One of the first puzzles we wanted to solve was the difficulty to push upgrades.
+In order to solve that, we designed 0-OS components as completely stand-alone modules. Each subsystem, be it storage, networking, containers/VMs, is managed by it's own component (mostly a daemon), and communicate with each-other through a local bus. And as we said, each component can then be upgraded separately, together with the necessary data migrations that could be required.
+
+**WHAT API?**
+
+The second big change is our approach to the API, or better, lack thereof.
+In V2 we dropped the idea to expose the primitives of the Node over an API.
+Instead, all the required knowledge to deploy workloads is directly embedded in 0-OS.
+So in order to have the node deploy a workload, we have created a blueprint like system where the user describes what his requirements in terms of compute power, storage and networking are, and the node applies that blueprint to make it reality.
+That approach has a few advantages:
+ - It greatly reduces the attack surface of the node because there is no more direct interaction between a user and a node.
+ - And it also allows us to have a greater control over how things are organized in the node itself. The node being its own boss, can decide to re-organize itself whenever needed to optimize the capacity it can provide.
+ - Having a blueprint with requirements, gives the grid the possibility to verify that blueprint on multiple levels before applying it. That is: as well on top level as on node level a blueprint can be verified for validity and signatures before any other action will be executed.
+
+**PING**
+
+The last major change is how we want to handle networking.
+The solution used during the lifetime of V1 exposed its limitations when we started scaling our networks to hundreds of nodes.
+So here again we started from scratch and created our own overlay network solution.
+That solution is based on the 'new kid on the block' in terms of VPN: [Wireguard](https://wireguard.io) and it's approach and usage will be fully explained in the next 0-OS article.
+For the eager ones of you, there are some specifications and also some documentation [here](https://github.com/threefoldtech/zosv2/tree/master/docs/network) and [there](https://github.com/threefoldtech/zosv2/tree/master/specs/network).
+
+## That's All, Folks (for now)
+So this little article as an intro to the brave new world of 0-OS.
+The Zero-OS team engages itself to regularly keep you updated on it's progress, the new features that will surely be added, and for the so inclined, add a lot more content for techies on how to actually use that novel beast.
+
+[Till next time](https://youtu.be/b9434BoGkNQ)
diff --git a/collections/documentation/developers/internals/zos/internals/identity/identity.md b/collections/documentation/developers/internals/zos/internals/identity/identity.md
new file mode 100644
index 0000000..9ed7400
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/identity/identity.md
@@ -0,0 +1,143 @@
+ Node ID Generation
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [ZBus](#zbus)
+- [Home Directory](#home-directory)
+- [Introduction](#introduction-1)
+- [On Node Booting](#on-node-booting)
+- [ID generation](#id-generation)
+- [Cryptography](#cryptography)
+ - [zinit unit](#zinit-unit)
+- [Interface](#interface)
+
+***
+
+## Introduction
+
+We explain the node ID generation process.
+
+## ZBus
+
+Identity module is available on zbus over the following channel
+
+| module | object | version |
+|--------|--------|---------|
+| identity|[manager](#interface)| 0.0.1|
+
+## Home Directory
+
+identity keeps some data in the following locations
+
+| directory | path|
+|----|---|
+| root| `/var/cache/modules/identity`|
+
+## Introduction
+
+Identity manager is responsible for maintaining the node identity (public key). The manager make sure the node has one valid ID during the entire lifetime of the node. It also provide service to sign, encrypt and decrypt data using the node identity.
+
+On first boot, the identity manager will generate an ID and then persist this ID for life.
+
+Since the identity daemon is the only one that can access the node private key, it provides an interface to sign, verify and encrypt data. This methods are available for other modules on the local node to use.
+
+## On Node Booting
+
+- Check if node already has a seed generated
+- If yes, load the node identity
+- If not, generate a new ID
+- Start the zbus daemon.
+
+## ID generation
+
+At this time of development the ID generated by identityd is the base58 encoded public key of a ed25519 key pair.
+
+The key pair itself is generated from a random seed of 32 bytes. It is this seed that is actually saved on the node. And during boot the key pair is re-generated from this seed if it exists.
+
+## Cryptography
+
+The signing and encryption capabilities of the identity module rely on this ed25519 key pair.
+
+For signing, it directly used the key pair.
+For public key encryption, the ed25519 key pair is converted to its cure25519 equivalent and then use use to encrypt the data.
+
+### zinit unit
+
+The zinit unit file of the module specify the command line, test command, and the order where the services need to be booted.
+
+`identityd` require `storaged` to make sure the seed is persisted over reboots, to make sure node has the same ID during the full life time of the node.
+The identityd daemon is only considered running if the seed file exists.
+
+```yaml
+exec: /bin/identityd
+test: test -e /var/cache/modules/identity/seed.txt
+after:
+ - storaged
+```
+
+## Interface
+
+For an up to date interface please check code [here](https://github.com/threefoldtech/zos/blob/main/pkg/identity.go)
+```go
+package pkg
+
+// Identifier is the interface that defines
+// how an object can be used as an identity
+type Identifier interface {
+ Identity() string
+}
+
+// StrIdentifier is a helper type that implement the Identifier interface
+// on top of simple string
+type StrIdentifier string
+
+// Identity implements the Identifier interface
+func (s StrIdentifier) Identity() string {
+ return string(s)
+}
+
+// IdentityManager interface.
+type IdentityManager interface {
+ // NodeID returns the node id (public key)
+ NodeID() StrIdentifier
+
+ // NodeIDNumeric returns the node registered ID.
+ NodeIDNumeric() (uint32, error)
+
+ // FarmID return the farm id this node is part of. this is usually a configuration
+ // that the node is booted with. An error is returned if the farmer id is not configured
+ FarmID() (FarmID, error)
+
+ // Farm returns name of the farm. Or error
+ Farm() (string, error)
+
+ //FarmSecret get the farm secret as defined in the boot params
+ FarmSecret() (string, error)
+
+ // Sign signs the message with privateKey and returns a signature.
+ Sign(message []byte) ([]byte, error)
+
+ // Verify reports whether sig is a valid signature of message by publicKey.
+ Verify(message, sig []byte) error
+
+ // Encrypt encrypts message with the public key of the node
+ Encrypt(message []byte) ([]byte, error)
+
+ // Decrypt decrypts message with the private of the node
+ Decrypt(message []byte) ([]byte, error)
+
+ // EncryptECDH aes encrypt msg using a shared key derived from private key of the node and public key of the other party using Elliptic curve Diffie Helman algorithm
+ // the nonce if prepended to the encrypted message
+ EncryptECDH(msg []byte, publicKey []byte) ([]byte, error)
+
+ // DecryptECDH decrypt aes encrypted msg using a shared key derived from private key of the node and public key of the other party using Elliptic curve Diffie Helman algorithm
+ DecryptECDH(msg []byte, publicKey []byte) ([]byte, error)
+
+ // PrivateKey sends the keypair
+ PrivateKey() []byte
+}
+
+// FarmID is the identification of a farm
+type FarmID uint32
+```
diff --git a/collections/documentation/developers/internals/zos/internals/identity/readme.md b/collections/documentation/developers/internals/zos/internals/identity/readme.md
new file mode 100644
index 0000000..1bff097
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/identity/readme.md
@@ -0,0 +1,8 @@
+ Identity Module
+
+Identity daemon is responsible for two major operations that are crucial for the node operation.
+
+ Table of Contents
+
+- [Node ID Generation](identity.md)
+- [Node Live Software Update](upgrade.md)
diff --git a/collections/documentation/developers/internals/zos/internals/identity/upgrade.md b/collections/documentation/developers/internals/zos/internals/identity/upgrade.md
new file mode 100644
index 0000000..e486a53
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/identity/upgrade.md
@@ -0,0 +1,98 @@
+ Node Upgrade
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Philosophy](#philosophy)
+- [Booting a new node](#booting-a-new-node)
+- [Runtime upgrade of a node](#runtime-upgrade-of-a-node)
+- [Technical](#technical)
+ - [Flist layout](#flist-layout)
+
+***
+
+## Introduction
+
+We provide information concerning node upgrade with ZOS. We also explain the philosophy behind ZOS.
+
+## Philosophy
+
+0-OS is meant to be a black box no one can access. While this provide some nice security features it also makes it harder to manage. Specially when it comes to update/upgrade.
+
+Hence, zos only trust few sources for upgrade packages. When the node boots up it checks the sources for the latest release and make sure all the local binaries are up-to-date before continuing the booting. The flist source must be rock-solid secured, that's another topic for different documentation.
+
+The run mode defines which flist the node is going to use to boot. Run mode can be specified by passing `runmode=` to the kernel boot params. Currently we have those different run modes.
+
+- dev: ephemeral network only setup to develop and test new features. Can be created and reset at anytime
+- test: Mostly stable features that need to be tested at scale, allow preview and test of new features. Always the latest and greatest. This network can be reset sometimes, but should be relatively stable.
+- prod: Released of stable version. Used to run the real grid with real money. Cannot be reset ever. Only stable and battle tested feature reach this level.
+
+## Booting a new node
+
+The base image for zos contains a very small subset of tools, plus the boot program. Standing alone, the image is not really useful. On boot and
+after initial start of the system, the boot program kicks in and it does the following:
+
+- Detect the boot flist that the node must use to fully start. The default is hard-coded into zos, but this can be overridden by the `flist=` kernel param. The `flist=` kernel param can get deprecated without a warning, since it's a development flag.
+- The bootstrap, will then mount this flist using 0-fs, this of course requires a working connection to the internet. Hence bootstrap is configured to wait for the `internet` service.
+- The flist information (name, and version) is saved under `/tmp/flist.name` and `/tmp/flist.info`.
+- The bootstrap makes sure to copy all files in the flist to the proper locations under the system rootfs, this include `zinit` config files.
+- Then zinit is asked to monitor new installed services, zinit takes care of those services and make sure they are properly working at all times.
+- Bootstrap, umounts the flist, cleans up before it exits.
+- Boot process continues.
+
+## Runtime upgrade of a node
+
+Once the node is up and running, identityd takes over and it does the following:
+
+- It loads the boot info files `/tmp/flist.name` and `/tmp/flist.info`
+- If the `flist.name` file does **not** exist, `identityd` will assume the node is booted with other means than an flist (for example overlay). In that case, identityd will log this, and disable live upgrade of the node.
+- If the `flist.name` file exists, the flist will be monitored on the `https://hub.grid.tf` for changes. Any change in the version will initiate a life upgrade routine.
+- Once the flist change is detected, identityd will mount the flist, make sure identityd is running the latest version. If not, identityd will update itself first before continuing.
+- services that will need update will be gracefully stopped.
+- `identityd` will then make sure to update all services from the flist, and config files. and restart the services properly.
+- services are started again after all binaries has been copied
+
+## Technical
+
+0-OS is designed to provide maximum uptime for its workload, rebooting a node should never be required to upgrade any of its component (except when we push a kernel upgrade).
+
+![flow](../../assets/0-OS-upgrade.png)
+
+### Flist layout
+
+The files in the upgrade flist needs to be located in the filesystem tree at the same destination they would need to be in 0-OS. This allow the upgrade code to stays simple and only does a copy from the flist to the root filesystem of the node.
+
+Booting a new node, or updating a node uses the same flist. Hence, a boot flist must container all required services for node operation.
+
+Example:
+
+0-OS filesystem:
+
+```
+/etc/zinit/identityd.yaml
+/etc/zinit/networkd.yaml
+/etc/zinit/contd.yaml
+/etc/zinit/init/node-ready.sh
+/etc/zinit/init
+/etc/zinit/redis.yaml
+/etc/zinit/storaged.yaml
+/etc/zinit/flistd.yaml
+/etc/zinit/readme.md
+/etc/zinit/internet.yaml
+/etc/zinit/containerd.yaml
+/etc/zinit/boot.yaml
+/etc/zinit/provisiond.yaml
+/etc/zinit/node-ready.yaml
+/etc/zinit
+/etc
+/bin/zlf
+/bin/provisiond
+/bin/flistd
+/bin/identityd
+/bin/contd
+/bin/capacityd
+/bin/storaged
+/bin/networkd
+/bin/internet
+/bin
+```
diff --git a/collections/documentation/developers/internals/zos/internals/internals.md b/collections/documentation/developers/internals/zos/internals/internals.md
new file mode 100644
index 0000000..ac7dfe8
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/internals.md
@@ -0,0 +1,88 @@
+ Internal Modules
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Booting](#booting)
+- [Bootstrap](#bootstrap)
+- [Zinit](#zinit)
+- [Architecture](#architecture)
+ - [IPC](#ipc)
+- [ZOS Processes (modules)](#zos-processes-modules)
+- [Capacity](#capacity)
+
+***
+
+## Introduction
+
+This document explains in a nutshell the internals of ZOS. This includes the boot process, architecture, the internal modules (and their responsibilities), and the inter-process communication.
+
+## Booting
+
+ZOS is a linux based operating system in the sense that we use the main-stream linux kernel with no modifications (but heavily customized). The base image of ZOS includes linux, busybox, [zinit](https://github.com/threefoldtech/zinit) and other required tools that are needed during the boot process. The base image is also shipped with a bootstrap utility that is self-updating on boot which kick starts everything.
+
+For more details about the ZOS base image please check [0-initramfs](https://github.com/threefoldtech/0-initramfs).
+
+`ZOS` uses zinit as its `init` or `PID 1` process. `zinit` acts as a process manager and it takes care of starting all required services in the right order. Using simple configuration that is available under `/etc/zinit`.
+
+The base `ZOS` image has a zinit config to start the basic services that are required for booting. These include (mainly) but are not limited to:
+
+- internet: A very basic service that tries to connect zos to the internet as fast (and as simple) as possible (over ethernet) using dhcp. This is needed so the system can continue the boot process. Once this one succeeds, it exits and leaves node network management to the more sophisticated ZOS module `networkd` which is yet to be downloaded and started by bootstrap.
+- redis: This is required by all zos modules for its IPC (inter process communication).
+- bootstrap: The bootstrap process which takes care of downloading all required zos binaries and modules. This one requires the `internet` service to actually succeed.
+
+## Bootstrap
+
+`bootstrap` is a utility that resides on the base image. It takes care of downloading and configuring all zos main services by doing the following:
+
+- It checks if there is a more recent version of itself available. If it exists, the process first updates itself before proceeding.
+- It checks zos boot parameters (for example, which network you are booting into) as set by .
+- Once the network is known, let's call it `${network}`. This can either be `production`, `testing`, or `development`. The proper release is downloaded as follows:
+ - All flists are downloaded from one of the [hub](https://hub.grid.tf/) `tf-zos-v3-bins.dev`, `tf-zos-v3-bins.test`, or `tf-zos-v3-bins` repos. Based on the network, only one of those repos is used to download all the support tools and binaries. Those are not included in the base image because they can be updated, added, or removed.
+ - The flist `https://hub.grid.tf/tf-zos/zos:${network}-3:latest.flist.md` is downloaded (note that ${network} is replaced with the actual value). This flist includes all zos services from this repository. More information about the zos modules are explained later.
+ - Once all binaries are downloaded, `bootstrap` finishes by asking zinit to start monitoring the newly installed services. The bootstrap exits and will never be started again as long as zos is running.
+ - If zos is restarted the entire bootstrap process happens again including downloading the binaries because ZOS is completely stateless (except for some cached runtime data that is preserved across reboots on a cache disk).
+
+## Zinit
+
+As mentioned earlier, `zinit` is the process manager of zos. Bootstrap makes sure it registers all zos services for zinit to monitor. This means that zinit will take care that those services are always running, and restart them if they have crashed for any reason.
+
+## Architecture
+
+For `ZOS` to be able to run workloads of different types it has split its functionality into smaller modules. Where each module is responsible for providing a single functionality. For example `storaged` which manages machine storages, hence it can provide low level storage capacity to other services that need it.
+
+As an example, imagine that you want to start a `virtual machine`. For a `virtual machine` to be able to run it will require a `rootfs` image or the image of the VM itself this is normally provided via an `flist` (managed by `flistd`), then you would need an actual persistent storage (managed by `storaged`), a virtual nic (managed by `networkd`), another service that can put everything together in a form of a VM (`vmd`). Then finally a service that orchestrates all of this and translates the user request to an actual workload `provisiond`, you get the picture.
+
+### IPC
+
+All modules running in zos needs to be able to interact with each other. As it shows from the previous example. For example, `provision` daemon need to be able to ask `storage` daemon to prepare a virtual disk. A new `inter-process communication` protocol and library was developed to enable this with those extra features:
+
+- Modules do not need to know where other modules live, there are no ports, and/or urls that have to be known by all services.
+- A single module can run multiple versions of an API.
+- Ease of development.
+- Auto generated clients.
+
+For more details about the message bus please check [zbus](https://github.com/threefoldtech/zbus)
+
+`zbus` uses redis as a message bus, hence redis is started in the early stages of zos booting.
+
+`zbus` allows auto generation of `stubs` which are generated clients against a certain module interface. Hence a module X can interact with a module Y by importing the generated clients and then start making function calls.
+
+## ZOS Processes (modules)
+
+Modules of zos are completely internal. There is no way for an external user to talk to them directly. The idea is the node exposes a public API over rmb, while internally this API can talk to internal modules over `zbus`.
+
+Here is a list of the major ZOS modules.
+
+- [Identity](identity/index.md)
+- [Node](node/index.md)
+- [Storage](storage/index.md)
+- [Network](network/index.md)
+- [Flist](flist/index.md)
+- [Container](container/index.md)
+- [VM](vmd/index.md)
+- [Provision](provision/index.md)
+
+## Capacity
+
+In [this document](./capacity.md), you can find detail description of how ZOS does capacity planning.
diff --git a/collections/documentation/developers/internals/zos/internals/macdev/readme.md b/collections/documentation/developers/internals/zos/internals/macdev/readme.md
new file mode 100644
index 0000000..1a9d4b6
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/macdev/readme.md
@@ -0,0 +1,57 @@
+> Note: This is unmaintained, try on your own responsibility
+
+# MacOS Developer
+
+0-OS (v2) uses a Linux kernel and is really build with a linux environment in mind.
+As a developer working from a MacOS environment you will have troubles running the 0-OS code.
+
+Using [Docker][docker] you can work from a Linux development environment, hosted from your MacOS Host machine.
+In this README we'll do exactly that using the standard Ubuntu [Docker][docker] container as our base.
+
+## Setup
+
+0. Make sure to have Docker installed, and configured (also make sure you have your code folder path shared in your Docker preferences).
+1. Start an _Ubuntu_ Docker container with your shared code directory mounted as a volume:
+```bash
+docker run -ti -v "$HOME/oss":/oss ubuntu /bin/bash
+```
+2. Make sure your environment is updated and upgraded using `apt-get`.
+3. Install Go (`1.13`) from src using the following link or the one you found on [the downloads page](https://golang.org/dl/):
+```bash
+wget https://dl.google.com/go/go1.13.3.linux-amd64.tar.gz
+sudo tar -xvf go1.13.3.linux-amd64.tar.gz
+sudo mv go /usr/local
+```
+4. Add the following to your `$HOME/.bashrc` and `source` it:
+```vim
+export GOROOT=/usr/local/go
+export GOPATH=$HOME/go
+export PATH=$GOPATH/bin:$GOROOT/bin:$PATH
+```
+5. Confirm you have Go installed correctly:
+```
+go version && go env
+```
+6. Go to your `zos` code `pkg` directory hosted from your MacOS development machine within your docker `/bin/bash`:
+```bash
+cd /oss/github.com/threefoldtech/zos/pkg
+```
+7. Install the dependencies for testing:
+```bash
+make getdeps
+```
+8. Run tests and verify all works as expected:
+```bash
+make test
+```
+9. Build `zos`:
+```bash
+make build
+```
+
+If you can successfully do step (8) and step (9) you
+can now contribute to `zos` as a MacOS developer.
+Testing and compiling you'll do from within your container's shell,
+coding you can do from your beloved IDE on your MacOS development environment.
+
+[docker]: https://www.docker.com
diff --git a/collections/documentation/developers/internals/zos/internals/network/Deploy_Network-V2.md b/collections/documentation/developers/internals/zos/internals/network/Deploy_Network-V2.md
new file mode 100644
index 0000000..59a04d9
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/network/Deploy_Network-V2.md
@@ -0,0 +1,74 @@
+# 0-OS v2 and it's network setup
+
+## Introduction
+
+0-OS nodes participating in the Threefold grid, need connectivity of course. They need to be able to communicate over
+the Internet with each-other in order to do various things:
+
+- download it's OS modules
+- perform OS module upgrades
+- register itself to the grid, and send regular updates about it's status
+- query the grid for tasks to execute
+- build and run the Overlay Network
+- download flists and the effective files to cache
+
+The nodes themselves can have connectivity in a few different ways:
+
+- Only have RFC1918 private addresses, connected to the Internet through NAT, NO IPv6
+ Mostly, these are single-NIC (Network card) machines that can host some workloads through the Overlay Network, but
+ cant't expose services directly. These are HIDDEN nodes, and are mostly booted with an USB stick from
+ bootstrap.grid.tf .
+- Dual-stacked: having RFC1918 private IPv4 and public IPv6 , where the IPv6 addresses are received from a home router,
+but firewalled for outgoing traffic only. These nodes are effectively also HIDDEN
+- Nodes with 2 NICs, one that has effectively a NIC connected to a segment that has real public
+addresses (IPv4 and/or IPv6) and one NIC that is used for booting and local
+management. (OOB) (like in the drawing for farmer setup)
+
+For Farmers, we need to have Nodes to be reachable over IPv6, so that the nodes can:
+
+- expose services to be proxied into containers/vms
+- act as aggregating nodes for Overlay Networks for HIDDEN Nodes
+
+Some Nodes in Farms should also have a publicly reachable IPv4, to make sure that clients that only have IPv4 can
+effectively reach exposed services.
+
+But we need to stress the importance of IPv6 availability when you're running a multi-node farm in a datacentre: as the
+grid is boldly claiming to be a new Internet, we should make sure we adhere to the new protocols that are future-proof.
+Hence: IPv6 is the base, and IPv4 is just there to accomodate the transition.
+
+Nowadays, RIPE can't even hand out consecutive /22 IPv4 blocks any more for new LIRs, so you'll be bound to market to
+get IPv4, mostly at rates of 10-15 Euro per IP. Things tend to get costly that way.
+
+So anyway, IPv6 is not an afterthought in 0-OS, we're starting with it.
+
+## Network setup for farmers
+
+This is a quick manual to what is needed for connecting a node with zero-OS V2.0
+
+### Step 1. Testing for IPv6 availability in your location
+As descibed above the network in which the node is instaleld has to be IPv6 enabled. This is not an afterthought as we are building a new internet it has to ba based on the new and forward looking IP addressing scheme. This is something you have to investigate, negotiate with you connectivity provider. Many (but not all home connectivity products and certainly most datacenters can provide you with IPv6. There are many sources of infromation on how to test and check whether your connection is IPv6 enabled, [here is a starting point](http://www.ipv6enabled.org/ipv6_enabled/ipv6_enable.php)
+
+### Step 2. Choosing you setup for connecitng you nodes.
+
+Once you have established that you have IPv6 enabled on the network you are about to deploy, you have to make sure that there is an IPv6 DHCP facility available. Zero-OS does not work with static IPv6 addresses (at this point in time). So you have choose and create one of the following setups:
+
+#### 2.1 Home setup
+
+Use your (home) ISP router Ipv6 DHCP capabilities to provide (private) IPv6 addresses. The principle will work the same as for IPv4 home connections, everything happens enabled by Network Adress Translation (just like anything else that uses internet connectivity). This should be relatively straightforward if you have established that your conenction has IPv6 enabled.
+
+#### 2.2 Datacenter / Expert setup
+
+In this situation there are many options on how to setup you node. This requires you as the expert to make a few decisions on how to connect what what the best setup is that you can support for the operaitonal time of your farm. The same basics principles apply:
+ - You have to have a block of (public) IPv6 routed to you router, or you have to have your router setup to provide Network Address Translation (NAT)
+ - You have to have a DHCP server in your network that manages and controls IPV6 ip adress leases. Depending on your specific setup you have this DHCP server manage a public IPv6y range which makes all nodes directly connected to the public internet or you have this DHCP server manage a private block og IPv6 addresses which makes all you nodes connect to the internet through NAT.
+
+As a farmer you are in charge of selecting and creating the appropriate network setup for your farm.
+
+## General notes
+
+The above setup will allows your node(s) to appear in explorer on the TF Grid and will allowd you to earn farming tokens. At stated in the introduction ThreeFold is creating next generation internet capacity and therefore has IPv6 as it's base building block. Connecting to the current (dominant) IPv4 network happens for IT workloads through so called webgateways. As the word sais these are gateways that provide connectivity between the currenct leading IPv4 adressing scheme and IPv6.
+
+We have started a forum where people share their experiences and configurations. This will be work in progress and forever growing.
+
+**IMPORTANT**: You as a farmer do not need access to IPV4 to be able to rent capacity for IT workloads that need to be visible on IPV4, this is something that can happen elswhere on the TF Grid.
+
diff --git a/collections/documentation/developers/internals/zos/internals/network/HIDDEN-PUBLIC.dia b/collections/documentation/developers/internals/zos/internals/network/HIDDEN-PUBLIC.dia
new file mode 100644
index 0000000..139cffa
Binary files /dev/null and b/collections/documentation/developers/internals/zos/internals/network/HIDDEN-PUBLIC.dia differ
diff --git a/collections/documentation/developers/internals/zos/internals/network/HIDDEN-PUBLIC.png b/collections/documentation/developers/internals/zos/internals/network/HIDDEN-PUBLIC.png
new file mode 100644
index 0000000..72fbe35
Binary files /dev/null and b/collections/documentation/developers/internals/zos/internals/network/HIDDEN-PUBLIC.png differ
diff --git a/collections/documentation/developers/internals/zos/internals/network/NR_layout.dia b/collections/documentation/developers/internals/zos/internals/network/NR_layout.dia
new file mode 100644
index 0000000..a9f59e2
Binary files /dev/null and b/collections/documentation/developers/internals/zos/internals/network/NR_layout.dia differ
diff --git a/collections/documentation/developers/internals/zos/internals/network/NR_layout.png b/collections/documentation/developers/internals/zos/internals/network/NR_layout.png
new file mode 100644
index 0000000..2336642
Binary files /dev/null and b/collections/documentation/developers/internals/zos/internals/network/NR_layout.png differ
diff --git a/collections/documentation/developers/internals/zos/internals/network/Network-V2.md b/collections/documentation/developers/internals/zos/internals/network/Network-V2.md
new file mode 100644
index 0000000..59e16ae
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/network/Network-V2.md
@@ -0,0 +1,315 @@
+# 0-OS v2 and it's network
+
+## Introduction
+
+0-OS nodes participating in the Threefold grid, need connectivity of course. They need to be able to communicate over
+the Internet with each-other in order to do various things:
+
+- download it's OS modules
+- perform OS module upgrades
+- register itself to the grid, and send regular updates about it's status
+- query the grid for tasks to execute
+- build and run the Overlay Network
+- download flists and the effective files to cache
+
+The nodes themselves can have connectivity in a few different ways:
+
+- Only have RFC1918 private addresses, connected to the Internet through NAT, NO IPv6
+ Mostly, these are single-NIC (Network card) machines that can host some workloads through the Overlay Network, but
+ cant't expose services directly. These are HIDDEN nodes, and are mostly booted with an USB stick from
+ bootstrap.grid.tf .
+- Dual-stacked: having RFC1918 private IPv4 and public IPv6 , where the IPv6 addresses are received from a home router,
+but firewalled for outgoing traffic only. These nodes are effectively also HIDDEN
+- Nodes with 2 NICs, one that has effectively a NIC connected to a segment that has real public
+addresses (IPv4 and/or IPv6) and one NIC that is used for booting and local
+management. (OOB) (like in the drawing for farmer setup)
+
+For Farmers, we need to have Nodes to be reachable over IPv6, so that the nodes can:
+
+- expose services to be proxied into containers/vms
+- act as aggregating nodes for Overlay Networks for HIDDEN Nodes
+
+Some Nodes in Farms should also have a publicly reachable IPv4, to make sure that clients that only have IPv4 can
+effectively reach exposed services.
+
+But we need to stress the importance of IPv6 availability when you're running a multi-node farm in a datacentre: as the
+grid is boldly claiming to be a new Internet, we should make sure we adhere to the new protocols that are future-proof.
+Hence: IPv6 is the base, and IPv4 is just there to accomodate the transition.
+
+Nowadays, RIPE can't even hand out consecutive /22 IPv4 blocks any more for new LIRs, so you'll be bound to market to
+get IPv4, mostly at rates of 10-15 Euro per IP. Things tend to get costly that way.
+
+So anyway, IPv6 is not an afterthought in 0-OS, we're starting with it.
+
+## Physical setup for farmers
+
+```text
+ XXXXX XXX
+ XX XXX XXXXX XXX
+ X X XXX
+ X X
+ X INTERNET X
+ XXX X X
+ XXXXX XX XX XXXX
+ +X XXXX XX XXXXX
+ |
+ |
+ |
+ |
+ |
+ +------+--------+
+ | FIREWALL/ |
+ | ROUTER |
+ +--+----------+-+
+ | |
+ +-----------+----+ +-+--------------+
+ | switch/ | | switch/ |
+ | vlan segment | | vlan segment |
+ +-+---------+----+ +---+------------+
+ | | |
++-------+-------+ |OOB | PUBLIC
+| PXE / dhcp | | |
+| Ser^er | | |
++---------------+ | |
+ | |
+ +-----+------------+----------+
+ | |
+ | +--+
+ | | |
+ | NODES | +--+
+ +--+--------------------------+ | |
+ | | |
+ +--+--------------------------+ |
+ | |
+ +-----------------------------+
+```
+
+The PXE/dhcp can also be done by the firewall, your mileage may vary.
+
+## Switch and firewall configs
+
+Single switch, multiple switch, it all boils down to the same:
+
+- one port is an access port on an OOB vlan/segment
+- one port is connected to a public vlan/segment
+
+The farmer makes sure that every node receives properly an IPv4 address in the OOB segment through means of dhcp, so
+that with a PXE config or USB, a node can effectively start it's boot process:
+
+- Download kernel and initrd
+- Download and mount the system flists so that the 0-OS daemons can start
+- Register itself on the grid
+- Query the grid for tasks to execute
+
+For the PUBLIC side of the Nodes, there are a few things to consider:
+
+- It's the farmer's job to inform the grid what node gets an IP address, be it IPv4 or IPv4.
+- Nodes that don't receive and IPv4 address will connect to the IPv4 net through the NATed OOB network
+- A farmer is responsible to provide and IPv6 prefix on at least one segment, and have a Router Advertisement daemon
+runnig to provide for SLAAC addressin on that segment.
+- That IPv6 Prefix on the public segment should not be firewalled, as it's impossible to know in your firewall what
+ports will get exposed for the proxies.
+
+The Nodes themselves have nothing listening that points into the host OS itself, and are by themselves also firewalled.
+In dev mode, there is an ssh server with a key-only login, accessible by a select few ;-)
+
+## DHCP/Radvd/RA/DHCP6
+
+For home networks, there is not much to do, a Node will get an IPv4 Private(rfc1918) address , and most probaly and
+ipv6 address in a /64 prefix, but is not reachable over ipv6, unless the firewall is disabled for IPv6. As we can't
+rely on the fact that that is possible, we assume these nodes to be HIDDEN.
+
+A normal self-respecting Firewall or IP-capable switch can hand out IP[46] addresses, some can
+even bootp/tftp to get nodes booted over the network.
+We are (full of hope) assuming that you would have such a beast to configure and splice your network
+in multiple segments.
+A segment is a physical network separation. That can be port-based vlans, or even separate switches, whatver rocks your
+boat, the keyword is here **separate**.
+
+On both segments you will need a way to hand out IPv4 addresses based on MAC addresses of the nodes. Yes, there is some
+administration to do, but it's a one-off, and really necessary, because you really need to know whic physical machine
+has which IP. For lights-out management and location of machines that is a must.
+
+So you'll need a list of mac addresses to add to your dhcp server for IPv4, to make sure you know which machine has
+received what IPv4 Address.
+That is necessary for 2 things:
+
+- locate the node if something is amiss, like be able to pinpoint a node's disk in case it broke (which it will)
+- have the node be reachable all the time, without the need to update the grid and network configs every time the node
+boots.
+
+## What happens under the hood (farmer)
+
+While we did our uttermost best to keep IPv4 address needs to a strict minimum, at least one Node will need an IPv4 address for handling everything that is Overlay Networks.
+For Containers to reach the Internet, any type of connectivity will do, be it NAT or though an Internal DMZ that has a
+routable IPv4 address.
+
+Internally, a lot of things are being set-up to have a node properly participate in the grid, as well to be prepared to partake in the User's Overlay Networks.
+
+A node connects itself to 'the Internet' depending on a few states.
+
+1. It lives in a fully private network (like it would be connected directly to a port on a home router)
+
+```
+ XX XXX
+ XXX XXXXXX
+ X Internet X
+ XXXXXXX XXXXX
+ XX XXX
+ XX X
+ X+X
+ |
+ |
+ +--------+-----------+
+ | HOME / |
+ | SOHO router |
+ | |
+ +--------+-----------+
+ |
+ | Private space IPv4
+ | (192.168.1.0/24)
+ |
++---------+------------+
+| |
+| NODE |
+| |
+| |
+| |
+| |
+| |
++----------------------+
+```
+
+1. It lives in a fully public network (like it is connected directly to an uplink and has a public ipv4 address)
+
+```
+ XX XXX
+ XXX XXXXXX
+ X Internet X
+ XXXXXXX XXXXX
+ XX XXX
+ XX X
+ X+X
+ |
+ | fully public space ipv4/6
+ | 185.69.166.0/24
+ | 2a02:1802:5e:0:1000::abcd/64
+ |
++---------+------------+
+| |
+| NODE |
+| |
++----------------------+
+
+```
+The node is fully reachable
+
+1. It lives in a datacentre, where a farmer manages the Network.
+
+A little Drawing :
+
+```text
++----------------------------------------------------+
+| switch |
+| |
+| |
++----------+-------------------------------------+---+
+ | |
+ access | |
+ mgmt | +---------------+
+ vlan | | access
+ | | public
+ | | vlan
+ | |
+ +-------+---------------------+------+
+ | |
+ | nic1 nic2 |
+ | |
+ | |
+ | |
+ | NODE |
+ | |
+ | |
+ | |
+ +------------------------------------+
+
+```
+
+Or the more elaborate drawing on top that should be sufficient for a sysadmin to comprehend.
+
+Although:
+
+- we don't (yet) support nic bonding (next release)
+- we don't (yet) support vlans, so your ports on switch/router need to be access ports to vlans to your router/firewall
+
+
+## yeayea, but really ... what now ?
+
+Ok, what are the constraints?
+
+A little foreword:
+ZosV2 uses IPv6 as it's base for networking, where the oldie IPv4 is merely an afterthought. So for it to work properly in it's actual incantation (we are working to get it to do IPv4-only too), for now, we need the node to live in a space that provides IPv6 __too__ .
+IPV4 and IPv6 are very different beasts, so any machine connected to the Internet wil do both on the same network. So basically your computer talks 2 different languages, when it comes to communicating. That is the same for ZOS, where right now, it's mother tongue is IPv6.
+
+So your zos for V2 can start in different settings
+1) you are a farmer, your ISP can provide you with IPv6
+Ok, you're all set, aside from a public IPv4 DHCP, you need to run a Stateless-Only SLAAC Router Advertiser (ZOS does NOT do DHCP6).
+
+1) you are a farmer, your ISP asks you what the hell IPv6 is
+That is problematic right now, wait for the next release of ZosV2
+
+1) you are a farmer, with only one node , at home, and on your PC https://ipv6.net tells you you have IPv6 on your PC.
+That means your home router received an IPV6 allocation from the ISP,
+Your'e all set, your node will boot, and register to the grid. If you know what you're doing, you can configure your router to allow all ipv6 traffic in forwarding mode to the specifice mac address of your node. (we'll explain later)
+1) you are a farmer, with a few nodes somewhere that are registered on the grid in V1, but you have no clue if IPv6 is supported where these nodes live
+1) you have a ThreefoldToken node at home, and still do not have a clue
+
+Basically it boils down also in a few other cases
+
+1) the physical network where a node lives has: IPv6 and Private space IPv4
+1) the physical network where a node lives has: IPv6 and Public IPv4
+1) the physical network where a node lives has: only IPv4
+
+But it bloils down to : call your ISP, ask for IPv6. It's the future, for yout ISP, it's time. There is no way to circumvent it. No way.
+
+
+OK, then, now what.
+
+1) you're a farmer with a bunch of nodes somewhere in a DC
+
+ - your nodes are connected once (with one NIC) to a switch/router
+ Then your router will have :
+ - a segment that carries IPv4 __and__ IPv6:
+
+ - for IPv4, there are 2 possibilities:
+ - it's RFC1918 (Private space) -> you NAT that subnet (e.g. 192.168.1.0/24) towards the Public Internet
+
+ - you __will__ have difficulty to designate a IPv4 public entrypoint into your farm
+ - your workloads will be only reachable through the overlay
+ - your storage will not be reachable
+
+ - you received a (small, because of the scarceness of IPv4 addresses, your ISP will give you only limited and pricy IPv4 adresses) IPv4 range you can utilise
+
+ - things are better, the nodes can live in public ipv4 space, where they can be used as entrypoint
+ - standard configuration that works
+
+ - for IPv6, your router is a Routing advertiser that provides SLAAC (Stateless, unmanaged) for that segment, working witha /64 prefix
+
+ - the nodes will reachable over IPv6
+ - storage backend will be available for the full grid
+ - everything will just work
+
+ Best solution for single NIC:
+ - an ipv6 prefx
+ - an ipv4 subnet (however small)
+
+ - your nodes have 2 connections, and you wnat to differ management from user traffic
+
+ - same applies as above, where the best outcome will be obtained with a real IPv6 prefix allocation and a small public subnet that is routable.
+ - the second NIC (typically 10GBit) will then carry everything public, and the first nic will just be there for managent, living in Private space for IPv4, mostly without IPv6
+ - your switch needs to be configured to provide port-based vlans, so the segments are properly separated, and your router needs to reflect that vlan config so that separation is handeled by the firewall in the router (iptables, pf, acl, ...)
+
+
+
+
+
diff --git a/collections/documentation/developers/internals/zos/internals/network/attic/exitpoints.md b/collections/documentation/developers/internals/zos/internals/network/attic/exitpoints.md
new file mode 100644
index 0000000..efecf3c
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/network/attic/exitpoints.md
@@ -0,0 +1,66 @@
+## Farmers providing transit for Tenant Networks (TN or Network)
+
+For networks of a user to be reachable, these networks need penultimate Network resources that act as exit nodes for the WireGuard mesh.
+
+For that Users need to sollicit a routable network with farmers that provide such a service.
+
+### Global registry for network resources. (`GRNR`?)
+
+Threefold through BCDB shoud keep a store where Farmers can register also a network service for Tenant Network (TN) reachablility.
+
+In a network transaction the first thing asked should be where a user wants to purchase it's transit. That can be with a nearby (latency or geolocation) Exit Provider (can e.g. be a Farmer), or with a Exit Provider outside of the geolocation for easier routing towards the primary entrypoint. (VPN-like services coming to mind)
+
+With this, we could envision in a later stage to have the Network Resources to be IPv6 multihomed with policy-based routing. That adds the possibiltiy to have multiple exit nodes for the same Network, with different IPv6 routes to them.
+
+### Datastructure
+
+A registered Farmer can also register his (dc-located?) network to be sold as transit space. For that he registers:
+ - the IPv4 addresses that can be allocated to exit nodes.
+ - the IPv6 prefix he obtained to be used in the Grid
+ - the nodes that will serve as exit nodes.
+ These nodes need to have IPv[46] access to routable address space through:
+ - Physical access in an interface of the node
+ - Access on a public `vlan` or via `vxlan / mpls / gre`
+
+Together with the registered nodes that will be part of that Public segment, the TNoDB (BCDB) can verify a Network Object containing an ExitPoint for a Network and add it to the queue for ExitNodes to fetch and apply.
+
+Physcally Nodes can be connected in several ways:
+ - living directly on the Internet (with a routable IPv4 and/or IPv6 Address) without Provider-enforced firewalling (outgoing traffic only)
+ - having an IPv4 allocation --and-- and IPv6 allocation
+ - having a single IPv4 address --and-- a single IPv6 allocation (/64) or even (Oh God Why) a single IPv6 addr.
+ - living in a Farm that has Nodes only reachable through NAT for IPv4 and no IPv6
+ - living in a Farm that has NAT IPv4 and routable IPv6 with an allocation
+ - living in a single-segment having IPv4 RFC1918 and only one IPv6 /64 prefix (home Nodes mostly)
+
+#### A Network resource allocation.
+We define Network Resource (NR) as a routable IPv6 `/64` Prefix, so for every time a new TNo is generated and validated, containing a new serial number and an added/removed NR, there has been a request to obtain a valid IPv6 Prefix (/64) to be added to the TNo.
+
+Basically it's just a list of allocations in that prefix, that are in use. Any free Prefix will do, as we do routing in the exit nodes with a `/64` granularity.
+
+The TNoDB (BCDB) then validates/updates the Tenant Network object with that new Network Resource and places it on a queue to be fetched by the interested Nodes.
+
+#### The Nodes responsible for ExitPoints
+
+A Node responsible for ExitPoints as wel as a Public endpoint will know so because of how it's registered in the TNoDB (BCDB). That is :
+ - it is defined as an exit node
+ - the TNoDB hands out an Object that describes it's public connectivity. i.e. :
+ - the public IPv4 address(es) it can use
+ - the IPv6 Prefix in the network segment that contains the penultimate default route
+ - an eventual Private BGP AS number for announcing the `/64` Prefixes of a Tenant Network, and the BGP peer(s).
+
+With that information, a Node can then build the Network Namespace from which it builds the Wireguard Interfaces prior to sending them in the ExitPoint Namespace.
+
+So the TNoDB (BCDB) hands out
+ - Tenant Network Objects
+ - Public Interface Objects
+
+They are related :
+ - A Node can have Network Resources
+ - A Network Resource can have (1) Public Interface
+ - Both are part of a Tenant Network
+
+A TNo defines a Network where ONLY the ExitPoint is flagged as being one. No more.
+When the Node (networkd) needs to setup a Public node, it will need to act differently.
+ - Verify if the Node is **really** public, if so use standard WG interface setup
+ - If not, verify if there is already a Public Exit Namespace defined, create WG interface there.
+ - If there is Public Exit Namespace, request one, and set it up first.
diff --git a/collections/documentation/developers/internals/zos/internals/network/attic/tools.md b/collections/documentation/developers/internals/zos/internals/network/attic/tools.md
new file mode 100644
index 0000000..8897bca
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/network/attic/tools.md
@@ -0,0 +1,264 @@
+# Network
+
+- [How does a farmer configure a node as exit node](#How-does-a-farmer-configure-a-node-as-exit-node)
+- [How to create a user private network](#How-to-create-a-user-private-network)
+
+## How does a farmer configure a node as exit node
+
+For the network of the grid to work properly, some of the nodes in the grid need to be configured as "exit nodes". An "exit node" is a node that has a publicly accessible IP address and that is responsible routing IPv6 traffic, or proxy IPv4 traffic.
+
+A farmer that wants to configure one of his nodes as "exit node", needs to register it in the TNODB. The node will then automatically detect it has been configured to be an exit node and do the necessary network configuration to start acting as one.
+
+At the current state of the development, we have a [TNODB mock](../../tools/tnodb_mock) server and a [tffarmer CLI](../../tools/tffarm) tool that can be used to do these configuration.
+
+Here is an example of how a farmer could register one of his node as "exit node":
+
+1. Farmer needs to create its farm identity
+
+```bash
+tffarmer register --seed myfarm.seed "mytestfarm"
+Farm registered successfully
+Name: mytestfarm
+Identity: ZF6jtCblLhTgAqp2jvxKkOxBgSSIlrRh1mRGiZaRr7E=
+```
+
+2. Boot your nodes with your farm identity specified in the kernel parameters.
+
+Take that farm identity create at step 1 and boot your node with the kernel parameters `farmer_id=`
+
+for your test farm that would be `farmer_id=ZF6jtCblLhTgAqp2jvxKkOxBgSSIlrRh1mRGiZaRr7E=`
+
+Once the node is booted, it will automatically register itself as being part of your farm into the [TNODB](../../tools/tnodb_mock) server.
+
+You can verify that you node registered itself properly by listing all the node from the TNODB by doing a GET request on the `/nodes` endpoints:
+
+```bash
+curl http://tnodb_addr/nodes
+[{"node_id":"kV3u7GJKWA7Js32LmNA5+G3A0WWnUG9h+5gnL6kr6lA=","farm_id":"ZF6jtCblLhTgAqp2jvxKkOxBgSSIlrRh1mRGiZaRr7E=","Ifaces":[]}]
+```
+
+3. Farmer needs to specify its public allocation range to the TNODB
+
+```bash
+tffarmer give-alloc 2a02:2788:0000::/32 --seed myfarm.seed
+prefix registered successfully
+```
+
+4. Configure the public interface of the exit node if needed
+
+In this step the farmer will tell his node how it needs to connect to the public internet. This configuration depends on the farm network setup, this is why this is up to the farmer to provide the detail on how the node needs to configure itself.
+
+In a first phase, we create the internet access in 2 ways:
+
+- the node is fully public: you don't need to configure a public interface, you can skip this step
+- the node has a management interface and a nic for public
+ then `configure-public` is required, and the farmer has the public interface connected to a specific public segment with a router to the internet in front.
+
+```bash
+tffarmer configure-public --ip 172.20.0.2/24 --gw 172.20.0.1 --iface eth1 kV3u7GJKWA7Js32LmNA5+G3A0WWnUG9h+5gnL6kr6lA=
+#public interface configured on node kV3u7GJKWA7Js32LmNA5+G3A0WWnUG9h+5gnL6kr6lA=
+```
+
+We still need to figure out a way to get the routes properly installed, we'll do static on the toplevel router for now to do a demo.
+
+The node is now configured to be used as an exit node.
+
+5. Mark a node as being an exit node
+
+The farmer then needs to select which node he agrees to use as an exit node for the grid
+
+```bash
+tffarmer select-exit kV3u7GJKWA7Js32LmNA5+G3A0WWnUG9h+5gnL6kr6lA=
+#Node kV3u7GJKWA7Js32LmNA5+G3A0WWnUG9h+5gnL6kr6lA= marked as exit node
+```
+
+## How to create a user private network
+
+1. Choose an exit node
+2. Request an new allocation from the farm of the exit node
+ - a GET request on the tnodb_mock at `/allocations/{farm_id}` will give you a new allocation
+3. Creates the network schema
+
+Steps 1 and 2 are easy enough to be done even manually but step 3 requires a deep knowledge of how networking works
+as well as the specific requirement of 0-OS network system.
+This is why we provide a tool that simplify this process for you, [tfuser](../../tools/tfuser).
+
+Using tfuser creating a network becomes trivial:
+
+```bash
+# creates a new network with node DLFF6CAshvyhCrpyTHq1dMd6QP6kFyhrVGegTgudk6xk as exit node
+# and output the result into network.json
+tfuser generate --schema network.json network create --node DLFF6CAshvyhCrpyTHq1dMd6QP6kFyhrVGegTgudk6xk
+```
+
+network.json will now contains something like:
+
+```json
+{
+ "id": "",
+ "tenant": "",
+ "reply-to": "",
+ "type": "network",
+ "data": {
+ "network_id": "J1UHHAizuCU6s9jPax1i1TUhUEQzWkKiPhBA452RagEp",
+ "resources": [
+ {
+ "node_id": {
+ "id": "DLFF6CAshvyhCrpyTHq1dMd6QP6kFyhrVGegTgudk6xk",
+ "farmer_id": "7koUE4nRbdsqEbtUVBhx3qvRqF58gfeHGMRGJxjqwfZi",
+ "reachability_v4": "public",
+ "reachability_v6": "public"
+ },
+ "prefix": "2001:b:a:8ac6::/64",
+ "link_local": "fe80::8ac6/64",
+ "peers": [
+ {
+ "type": "wireguard",
+ "prefix": "2001:b:a:8ac6::/64",
+ "Connection": {
+ "ip": "2a02:1802:5e::223",
+ "port": 1600,
+ "key": "PK1L7n+5Fo1znwD/Dt9lAupL19i7a6zzDopaEY7uOUE=",
+ "private_key": "9220e4e29f0acbf3bd7ef500645b78ae64b688399eb0e9e4e7e803afc4dd72418a1c5196208cb147308d7faf1212758042f19f06f64bad6ffe1f5ed707142dc8cc0a67130b9124db521e3a65e4aee18a0abf00b6f57dd59829f59662"
+ }
+ }
+ ],
+ "exit_point": true
+ }
+ ],
+ "prefix_zero": "2001:b:a::/64",
+ "exit_point": {
+ "ipv4_conf": null,
+ "ipv4_dnat": null,
+ "ipv6_conf": {
+ "addr": "fe80::8ac6/64",
+ "gateway": "fe80::1",
+ "metric": 0,
+ "iface": "public"
+ },
+ "ipv6_allow": []
+ },
+ "allocation_nr": 0,
+ "version": 0
+ }
+}
+```
+
+Which is a valid network schema. This network only contains a single exit node though, so not really useful.
+Let's add another node to the network:
+
+```bash
+tfuser generate --schema network.json network add-node --node 4hpUjrbYS4YeFbvLoeSR8LGJKVkB97JyS83UEhFUU3S4
+```
+
+result looks like:
+
+```json
+{
+ "id": "",
+ "tenant": "",
+ "reply-to": "",
+ "type": "network",
+ "data": {
+ "network_id": "J1UHHAizuCU6s9jPax1i1TUhUEQzWkKiPhBA452RagEp",
+ "resources": [
+ {
+ "node_id": {
+ "id": "DLFF6CAshvyhCrpyTHq1dMd6QP6kFyhrVGegTgudk6xk",
+ "farmer_id": "7koUE4nRbdsqEbtUVBhx3qvRqF58gfeHGMRGJxjqwfZi",
+ "reachability_v4": "public",
+ "reachability_v6": "public"
+ },
+ "prefix": "2001:b:a:8ac6::/64",
+ "link_local": "fe80::8ac6/64",
+ "peers": [
+ {
+ "type": "wireguard",
+ "prefix": "2001:b:a:8ac6::/64",
+ "Connection": {
+ "ip": "2a02:1802:5e::223",
+ "port": 1600,
+ "key": "PK1L7n+5Fo1znwD/Dt9lAupL19i7a6zzDopaEY7uOUE=",
+ "private_key": "9220e4e29f0acbf3bd7ef500645b78ae64b688399eb0e9e4e7e803afc4dd72418a1c5196208cb147308d7faf1212758042f19f06f64bad6ffe1f5ed707142dc8cc0a67130b9124db521e3a65e4aee18a0abf00b6f57dd59829f59662"
+ }
+ },
+ {
+ "type": "wireguard",
+ "prefix": "2001:b:a:b744::/64",
+ "Connection": {
+ "ip": "",
+ "port": 0,
+ "key": "3auHJw3XHFBiaI34C9pB/rmbomW3yQlItLD4YSzRvwc=",
+ "private_key": "96dc64ff11d05e8860272b91bf09d52d306b8ad71e5c010c0ccbcc8d8d8f602c57a30e786d0299731b86908382e4ea5a82f15b41ebe6ce09a61cfb8373d2024c55786be3ecad21fe0ee100339b5fa904961fbbbd25699198c1da86c5"
+ }
+ }
+ ],
+ "exit_point": true
+ },
+ {
+ "node_id": {
+ "id": "4hpUjrbYS4YeFbvLoeSR8LGJKVkB97JyS83UEhFUU3S4",
+ "farmer_id": "7koUE4nRbdsqEbtUVBhx3qvRqF58gfeHGMRGJxjqwfZi",
+ "reachability_v4": "hidden",
+ "reachability_v6": "hidden"
+ },
+ "prefix": "2001:b:a:b744::/64",
+ "link_local": "fe80::b744/64",
+ "peers": [
+ {
+ "type": "wireguard",
+ "prefix": "2001:b:a:8ac6::/64",
+ "Connection": {
+ "ip": "2a02:1802:5e::223",
+ "port": 1600,
+ "key": "PK1L7n+5Fo1znwD/Dt9lAupL19i7a6zzDopaEY7uOUE=",
+ "private_key": "9220e4e29f0acbf3bd7ef500645b78ae64b688399eb0e9e4e7e803afc4dd72418a1c5196208cb147308d7faf1212758042f19f06f64bad6ffe1f5ed707142dc8cc0a67130b9124db521e3a65e4aee18a0abf00b6f57dd59829f59662"
+ }
+ },
+ {
+ "type": "wireguard",
+ "prefix": "2001:b:a:b744::/64",
+ "Connection": {
+ "ip": "",
+ "port": 0,
+ "key": "3auHJw3XHFBiaI34C9pB/rmbomW3yQlItLD4YSzRvwc=",
+ "private_key": "96dc64ff11d05e8860272b91bf09d52d306b8ad71e5c010c0ccbcc8d8d8f602c57a30e786d0299731b86908382e4ea5a82f15b41ebe6ce09a61cfb8373d2024c55786be3ecad21fe0ee100339b5fa904961fbbbd25699198c1da86c5"
+ }
+ }
+ ],
+ "exit_point": false
+ }
+ ],
+ "prefix_zero": "2001:b:a::/64",
+ "exit_point": {
+ "ipv4_conf": null,
+ "ipv4_dnat": null,
+ "ipv6_conf": {
+ "addr": "fe80::8ac6/64",
+ "gateway": "fe80::1",
+ "metric": 0,
+ "iface": "public"
+ },
+ "ipv6_allow": []
+ },
+ "allocation_nr": 0,
+ "version": 1
+ }
+}
+```
+
+Our network schema is now ready, but before we can provision it onto a node, we need to sign it and send it to the bcdb.
+To be able to sign it we need to have a pair of key. You can use `tfuser id` command to create an identity:
+
+```bash
+tfuser id --output user.seed
+```
+
+We can now provision the network on both nodes:
+
+```bash
+tfuser provision --schema network.json \
+--node DLFF6CAshvyhCrpyTHq1dMd6QP6kFyhrVGegTgudk6xk \
+--node 4hpUjrbYS4YeFbvLoeSR8LGJKVkB97JyS83UEhFUU3S4 \
+--seed user.seed
+```
diff --git a/collections/documentation/developers/internals/zos/internals/network/attic/zostst.dhcp b/collections/documentation/developers/internals/zos/internals/network/attic/zostst.dhcp
new file mode 100644
index 0000000..0ac53be
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/network/attic/zostst.dhcp
@@ -0,0 +1,54 @@
+#!/usr/bin/bash
+
+mgmtnic=(
+0c:c4:7a:51:e3:6a
+0c:c4:7a:51:e9:e6
+0c:c4:7a:51:ea:18
+0c:c4:7a:51:e3:78
+0c:c4:7a:51:e7:f8
+0c:c4:7a:51:e8:ba
+0c:c4:7a:51:e8:0c
+0c:c4:7a:51:e7:fa
+)
+
+ipminic=(
+0c:c4:7a:4c:f3:b6
+0c:c4:7a:4d:02:8c
+0c:c4:7a:4d:02:91
+0c:c4:7a:4d:02:62
+0c:c4:7a:4c:f3:7e
+0c:c4:7a:4d:02:98
+0c:c4:7a:4d:02:19
+0c:c4:7a:4c:f2:e0
+)
+cnt=1
+for i in ${mgmtnic[*]} ; do
+cat << EOF
+config host
+ option name 'zosv2tst-${cnt}'
+ option dns '1'
+ option mac '${i}'
+ option ip '10.5.0.$((${cnt} + 10))'
+
+EOF
+let cnt++
+done
+
+
+
+cnt=1
+for i in ${ipminic[*]} ; do
+cat << EOF
+config host
+ option name 'ipmiv2tst-${cnt}'
+ option dns '1'
+ option mac '${i}'
+ option ip '10.5.0.$((${cnt} + 100))'
+
+EOF
+let cnt++
+done
+
+for i in ${mgmtnic[*]} ; do
+ echo ln -s zoststconf 01-$(echo $i | sed s/:/-/g)
+done
diff --git a/collections/documentation/developers/internals/zos/internals/network/definitions.md b/collections/documentation/developers/internals/zos/internals/network/definitions.md
new file mode 100644
index 0000000..137fe0a
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/network/definitions.md
@@ -0,0 +1,35 @@
+ Definitions
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Node](#node)
+- [TNo : Tenant Network Object](#tno--tenant-network-object)
+- [NR: Network Resource](#nr-network-resource)
+
+***
+
+## Introduction
+
+We present definitions of words used through the documentation.
+
+## Node
+
+ TL;DR: Computer.
+ A Node is a computer with CPU, Memory, Disks (or SSD's, NVMe) connected to _A_ network that has Internet access. (i.e. it can reach www.google.com, just like you on your phone, at home)
+ That Node will, once it has received an IP address (IPv4 or IPv6), register itself when it's new, or confirm it's identity and it's online-ness (for lack of a better word).
+
+## TNo : Tenant Network Object
+
+ TL;DR: The Network Description.
+ We named it so, because it is a data structure that describes the __whole__ network a user can request (or setup).
+ That network is a virtualized overlay network.
+ Basically that means that transfer of data in that network *always* is encrypted, protected from prying eyes, and __resources in that network can only communicate with each other__ **unless** there is a special rule that allows access. Be it by allowing access through firewall rules, *and/or* through a proxy (a service that forwards requests on behalf of, and ships replies back to the client).
+
+## NR: Network Resource
+
+ TL;DR: the Node-local part of a TNo.
+ The main building block of a TNo; i.e. each service of a user in a Node lives in an NR.
+ Each Node hosts User services, whatever type of service that is. Every service in that specific node will always be solely part of the Tenant's Network. (read that twice).
+ So: A Network Resource is the thing that interconnects all other network resources of the TN (Tenant Network), and provides routing/firewalling for these interconnects, including the default route to the BBI (Big Bad Internet), aka ExitPoint.
+ All User services that run in a Node are in some way or another connected to the Network Resource (NR), which will provide ip packet forwarding and firewalling to all other network resources (including the Exitpoint) of the TN (Tenant Network) of the user. (read that three times, and the last time, read it slowly and out loud)
\ No newline at end of file
diff --git a/collections/documentation/developers/internals/zos/internals/network/introduction.md b/collections/documentation/developers/internals/zos/internals/network/introduction.md
new file mode 100644
index 0000000..76660fa
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/network/introduction.md
@@ -0,0 +1,87 @@
+ Introduction to Networkd
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Boot and initial setup](#boot-and-initial-setup)
+- [Networkd functionality](#networkd-functionality)
+- [Techie talk](#techie-talk)
+- [Wireguard explanations](#wireguard-explanations)
+- [Caveats](#caveats)
+
+***
+
+## Introduction
+
+We provide an introduction to Networkd, the network manager of 0-OS.
+
+## Boot and initial setup
+
+At boot, be it from an usb stick or PXE, ZOS starts up the kernel, with a few necessary parameters like farm ID and/or possible network parameters, but basically once the kernel has started, [zinit](https://github.com/threefoldtech/zinit) among other things, starts the network initializer.
+
+In short, that process loops over the available network interfaces and tries to obtain an IP address that also provides for a default gateway. That means: it tries to get Internet connectivity. Without it, ZOS stops there, as not being able to register itself, nor start other processes, there wouldn't be any use for it to be started anyway.
+
+Once it has obtained Internet connectivity, ZOS can then proceed to make itself known to the Grid, and acknowledge it's existence. It will then regularly poll the Grid for tasks.
+
+Once initialized, with the network daemon running (a process that will handle all things related to networking), ZOS will set up some basic services so that workloads can themselves use that network.
+
+## Networkd functionality
+
+The network daemon is in itself responsible for a few tasks, and working together with the [provision daemon](../provision) it mainly sets up the local infrastructure to get the user network resources, together with the wireguard configurations for the user's mesh network.
+
+The Wireguard mesh is an overlay network. That means that traffic of that network is encrypted and encapsulated in a new traffic frame that the gets transferred over the underlay network, here in essence the network that has been set up during boot of the node.
+
+For users or workloads that run on top of the mesh, the mesh network looks and behaves like any other directly connected workload, and as such that workload can reach other workloads or services in that mesh with the added advantage that that traffic is encrypted, protecting services and communications over that mesh from too curious eyes.
+
+That also means that workloads between nodes in a local network of a farmer is even protected from the farmer himself, in essence protecting the user from the farmer in case that farmer could become too curious.
+
+As the nodes do not have any way to be accessed, be it over the underlaying network or even the local console of the node, a user can be sure that his workload cannot be snooped upon.
+
+## Techie talk
+
+- **boot and initial setup**
+For ZOS to work at all (the network is the computer), it needs an internet connection. That is: it needs to be able to communicate with the BCDB over the internet.
+So ZOS starts with that: with the `internet` process, that tries go get the node to receive an IP address. That process will have set-up a bridge (`zos`), connected to an interface that is on an Internet-capable network. That bridge will have an IP address that has Internet access.
+Also, that bridge is there for future public interfaces into workloads.
+Once ZOS can reach the Internet, the rest of the system can be started, where ultimately, the `networkd` daemon is started.
+
+- **networkd initial setup**
+`networkd` starts with recensing the available Network interfaces, and registers them to the BCDB (grid database), so that farmers can specify non-standard configs like for multi-nic machines. Once that is done, `networkd` registers itself to the zbus, so it can receive tasks to execute from the provsioning daemon (`provisiond`).
+These tasks are mostly setting up network resources for users, where a network resource is a subnet in the user's wireguard mesh.
+
+- **multi-nic setups**
+
+When someone is a farmer, exploiting nodes somewhere in a datacentre, where the nodes have multiple NICs, it is advisable (though not necessary) to differentiate OOB traffic (like initial boot setup) from user traffic (as well the overlay network as the outgoing NAT for nodes for IPv4) to be on a different NIC. With these parameters, a user will have to make sure their switches are properly configured, more in docs later.
+
+- **registering and configurations**
+
+Once a node has booted and properly initialized, registering and configuring the node to be able to accept workloads and their associated network configs, is a two-step process.
+First, the node registers it's live network setup to the BCDB. That is : all NICs with their associated IP addresses and routes are registered so a farm admin can in a second phase configure eventual separate NICs to handle different kinds of workloads.
+In that secondary phase, a farm admin can then set-up the NICs and their associated IP's manually, so that workloads can start using them.
+
+## Wireguard explanations
+
+- **wireguard as pointopoint links and what that means**
+Wireguard is a special type of VPN, where every instance is as well server for multiple peers as client towards multiple peers. That way you can create fanning-out connections als receive connections from multiple peers, creating effectively a mesh of connections Like this : ![like so](HIDDEN-PUBLIC.png)
+
+- **wireguard port management**
+Every wireguard point (a network resource point) needs a destination/port combo when it's publicly reachable. The destination is a public ip, but the port is the differentiator. So we need to make sure every network wireguard listening port is unique in the node where it runs, and can be reapplied in case of a node's reboot.
+ZOS registers the ports **already in use** to the BCDB, so a user can the pick a port that is not yet used.
+
+- **wireguard and hidden nodes**
+Hidden nodes are nodes that are in essence hidden behind a firewall, and unreachable from the Internet to an internal network, be it as an IPv4 NATed host or an IPv6 host that is firewalled in any way, where it's impossible to have connection initiations form the Internet to the node.
+As such, these nodes can only partake in a network as client-only towards publicly reachable peers, and can only initiate the connections themselves. (ref previous drawing).
+To make sure connectivity stays up, the clients (all) have a keepalive towards all their peers so that communications towards network resources in hidden nodes can be established.
+
+## Caveats
+
+- **hidden nodes**
+Hidden nodes live (mostly) behind firewalls that keep state about connections and these states have a lifetime. We try at best to keep these communications going, but depending of the firewall your mileage may vary (YMMV ;-))
+
+- **local underlay network reachability**
+When multiple nodes live in a same hidden network, at the moment we don't try to have the nodes establish connectivity between themselves, so all nodes in that hidden network can only reach each other through the intermediary of a node that is publicly reachable. So to get some performance, a farmer will have to have real routable nodes available in the vicinity.
+So for now, a farmer is better off to have his nodes really reachable over a public network.
+
+- **IPv6 and IPv4 considerations**
+While the mesh can work over IPv4 __and__ IPv6 at the same time, the peers can only be reached through one protocol at the same time. That is a peer is IPv4 __or__ IPv6, not both. Hence if a peer is reachable over IPv4, the client towards that peer needs to reach it over IPv4 too and thus needs an IPv4 address.
+We advise strongly to have all nodes properly set-up on a routable unfirewalled IPv6 network, so that these problems have no reason to exist.
diff --git a/collections/documentation/developers/internals/zos/internals/network/mesh.md b/collections/documentation/developers/internals/zos/internals/network/mesh.md
new file mode 100644
index 0000000..fd9bb85
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/network/mesh.md
@@ -0,0 +1,134 @@
+ Zero-Mesh
+
+ Table of Contents
+
+- [What It Is](#what-it-is)
+- [Overlay Network](#overlay-network)
+- [ZOS networkd](#zos-networkd)
+- [Internet reachability per Network Resource](#internet-reachability-per-network-resource)
+- [Interworkings](#interworkings)
+- [Network Resource Internals](#network-resource-internals)
+
+***
+
+## What It Is
+
+When a user wants to deploy a workload, whatever that may be, that workload needs connectivity.
+If there is just one service to be run, things can be simple, but in general there are more than one services that need to interact to provide a full stack. Sometimes these services can live on one node, but mostly these service will be deployed over multiple nodes, in different containers.
+The Mesh is created for that, where containers can communicate over an encrypted path, and that network can be specified in terms of IP addresses by the user.
+
+## Overlay Network
+
+Zero-Mesh is an overlay network. That requires that nodes need a proper working network with existing access to the Internet in the first place, being full-blown public access, or behind a firewall/home router that provides for Private IP NAT to the internet.
+
+Right now Zero-Mesh has support for both, where nodes behind a firewall are HIDDEN nodes, and nodes that are directly connected, be it over IPv6 or IPv4 as 'normal' nodes.
+Hidden nodes can thus only be participating as client nodes for a specific user Mesh, and all publicly reachable nodes can act as aggregators for hidden clients in that user Mesh.
+
+Also, a Mesh is static: once it is configured, and thus during the lifetime of the network, there is one node containing the aggregator for Mesh clients that live on hidden nodes. So if then an aggregator node has died or is not reachable any more, the mesh needs to be reapplied, with __some__ publicly reachable node as aggregator node.
+
+So it goes a bit like ![this](HIDDEN-PUBLIC.png)
+The Exit labeled NR in that graph is the point where Network Resources in Hidden Nodes connect to. These Exit NRs are then the transfer nodes between Hidden NRs.
+
+## ZOS networkd
+
+The networkd daemon receives tasks from the provisioning daemon, so that it can create the necessary resources for a Mesh participator in the User Network (A network Resource - NR).
+
+A network is defined as a whole by the User, using the tools in the 3bot to generate a proper configuration that can be used by the network daemon.
+
+What networkd takes care of, is the establishment of the mesh itself, in accordance with the configuration a farmer has given to his nodes. What is configured on top of the Mesh is user defined, and applied as such by the networkd.
+
+## Internet reachability per Network Resource
+
+Every node that participates in a User mesh, will also provide for Internet access for every network resource.
+that means that every NR has the same Internet access as the node itself. Which also means, in terms of security, that a firewall in the node takes care of blocking all types of entry to the NR, effectively being an Internet access diode, for outgoing and related traffic only.
+In a later phase a user will be able to define some network resource as __sole__ outgoing Internet Access point, but for now that is not yet defined.
+
+## Interworkings
+
+So How is that set up ?
+
+Every node participating in a User Network, sets up a Network Resource.
+Basically, it's a Linux Network Namespace (sort of a network virtual machine), that contains a wireguard interface that has a list of other Network resources it needs to route encrypted packets toward.
+
+As a User Network has a range typically a `/16` (like `10.1.0.0/16`), that is user defined. The User then picks a subnet from that range (like e.g. `10.1.1.0/24`) to assign that to every new NR he wants to participate in that Network.
+
+Workloads that are then provisioned are started in a newly created Container, and that container gets a User assigned IP __in__ that subnet of the Network Resource.
+
+The Network resource itself then handles the routing and firewalling for the containers that are connected to it. Also, the Network Resource takes care of internet connectivity, so that the container can reach out to other services on the Internet.
+
+![like this](NR_layout.png)
+
+Also in a later phase, a User will be able to add IPv6 prefixes to his Network Resources, so that containers are reachable over IPv6.
+
+Fully-routed IPv6 will then be available, where an Exit NR will be the entrypoint towards that network.
+
+## Network Resource Internals
+
+Each NR is basically a router for the User Network, but to allow NRs to access the Internet through the Node's local connection, there are some other internal routers to be added.
+
+Internally it looks like this :
+
+```text
++------------------------------------------------------------------------------+
+| |wg mesh |
+| +-------------+ +-----+-------+ |
+| | | | NR cust1 | 100.64.0.123/16 |
+| | container +----------+ 10.3.1.0/24 +----------------------+ |
+| | cust1 | veth| | public | |
+| +-------------+ +-------------+ | |
+| | |
+| +-------------+ +-------------+ | |
+| | | | NR cust200 | 100.64.4.200/16 | |
+| | container +----------+ 10.3.1.0/24 +----------------------+ |
+| | cust200 | veth| | public | |
+| +-------------+ +------+------+ | |
+| |wg mesh | |
+| 10.101.123.34/16 | |
+| +------------+ |tonrs |
+| | | +------------------+ |
+| | zos +------+ | 100.64.0.1/16 | |
+| | | | 10.101.12.231/16| ndmz | |
+| +---+--------+ NIC +-----------------------------+ | |
+| | | public +------------------+ |
+| +--------+------+ |
+| | |
+| | |
++------------------------------------------------------------------------------+
+ |
+ |
+ |
+ | 10.101.0.0/16 10.101.0.1
+ +------------------+------------------------------------------------------------
+
+ NAT
+ --------
+ rules NR custA
+ nft add rule inet nat postrouting oifname public masquerade
+ nft add rule inet filter input iifname public ct state { established, related } accept
+ nft add rule inet filter input iifname public drop
+
+ rules NR custB
+ nft add rule inet nat postrouting oifname public masquerade
+ nft add rule inet filter input iifname public ct state { established, related } accept
+ nft add rule inet filter input iifname public drop
+
+ rules ndmz
+ nft add rule inet nat postrouting oifname public masquerade
+ nft add rule inet filter input iifname public ct state { established, related } accept
+ nft add rule inet filter input iifname public drop
+
+
+ Routing
+
+ if NR only needs to get out:
+ ip route add default via 100.64.0.1 dev public
+
+ if an NR wants to use another NR as exitpoint
+ ip route add default via destnr
+ with for AllowedIPs 0.0.0.0/0 on that wg peer
+
+```
+
+During startup of the Node, the ndmz is put in place, following the configuration if it has a single internet connection , or that with a dual-nic setup, a separate nic is used for internet access.
+
+The ndmz network has the carrier-grade nat allocation assigned, so we don'tinterfere with RFC1918 private IPv4 address space, so users can use any of them (and not any of `100.64.0.0/10`, of course)
diff --git a/collections/documentation/developers/internals/zos/internals/network/readme.md b/collections/documentation/developers/internals/zos/internals/network/readme.md
new file mode 100644
index 0000000..3958cf6
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/network/readme.md
@@ -0,0 +1,8 @@
+ Zero-OS Networking
+
+ Table of Contents
+
+- [Introduction to networkd](./introduction.md)
+- [Vocabulary Definitions](./definitions.md)
+- [Wireguard Mesh Details](./mesh.md)
+- [Farm Network Setup](./setup_farm_network.md)
\ No newline at end of file
diff --git a/collections/documentation/developers/internals/zos/internals/network/setup_farm_network.md b/collections/documentation/developers/internals/zos/internals/network/setup_farm_network.md
new file mode 100644
index 0000000..23ecc84
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/network/setup_farm_network.md
@@ -0,0 +1,123 @@
+Setup
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Running ZOS (v2) at home](#running-zos-v2-at-home)
+- [Running ZOS (v2) in a multi-node farm in a DC](#running-zos-v2-in-a-multi-node-farm-in-a-dc)
+ - [Necessities](#necessities)
+ - [IPv6](#ipv6)
+ - [Routing/firewalling](#routingfirewalling)
+ - [Multi-NIC Nodes](#multi-nic-nodes)
+ - [Farmers and the grid](#farmers-and-the-grid)
+
+***
+
+## Introduction
+
+We present ZOSv2 network considerations.
+
+Running ZOS on a node is just a matter of booting it with a USB stick, or with a dhcp/bootp/tftp server with the right configuration so that the node can start the OS.
+Once it starts booting, the OS detects the NICs, and starts the network configuration. A Node can only continue it's boot process till the end when it effectively has received an IP address and a route to the Internet. Without that, the Node will retry indefinitely to obtain Internet access and not finish it's startup.
+
+So a Node needs to be connected to a __wired__ network, providing a dhcp server and a default gateway to the Internet, be it NATed or plainly on the public network, where any route to the Internet, be it IPv4 or IPv6 or both is sufficient.
+
+For a node to have that ability to host user networks, we **strongly** advise to have a working IPv6 setup, as that is the primary IP stack we're using for the User Network's Mesh to function.
+
+## Running ZOS (v2) at home
+
+Running a ZOS Node at home is plain simple. Connect it to your router, plug it in the network, insert the preconfigured USB stick containing the bootloader and the `farmer_id`, power it on.
+You will then see it appear in the Cockpit (`https://cockpit.testnet.grid.tf/capacity`), under your farm.
+
+## Running ZOS (v2) in a multi-node farm in a DC
+
+Multi-Node Farms, where a farmer wants to host the nodes in a data centre, have basically the same simplicity, but the nodes can boot from a boot server that provides for DHCP, and also delivers the iPXE image to load, without the need for a USB stick in every Node.
+
+A boot server is not really necessary, but it helps ;-). That server has a list of the MAC addresses of the nodes, and delivers the bootloader over PXE. The farmer is responsible to set-up the network, and configure the boot server.
+
+### Necessities
+
+The Farmer needs to:
+
+- Obtain an IPv6 prefix allocation from the provider. A `/64` will do, that is publicly reachable, but a `/48` is advisable if the farmer wants to provide IPv6 transit for User Networks
+- If IPv6 is not an option, obtain an IPv4 subnet from the provider. At least one IPv4 address per node is needed, where all IP addresses are publicly reachable.
+- Have the Nodes connected on that public network with a switch so that all Nodes are publicly reachable.
+- In case of multiple NICS, also make sure his farm is properly registered in BCDB, so that the Node's public IP Addresses are registered.
+- Properly list the MAC addresses of the Nodes, and configure the DHCP server to provide for an IP address, and in case of multiple NICs also provide for private IP addresses over DHCP per Node.
+- Make sure that after first boot, the Nodes are reachable.
+
+### IPv6
+
+IPv6, although already a real protocol since '98, has seen reluctant adoption over the time it exists. That mostly because ISPs and Carriers were reluctant to deploy it, and not seeing the need since the advent of NAT and private IP space, giving the false impression of security.
+But this month (10/2019), RIPE sent a mail to all it's LIRs that the last consecutive /22 in IPv4 has been allocated. Needless to say, but that makes the transition to IPv6 in 2019 of utmost importance and necessity.
+Hence, ZOS starts with IPv6, and IPv4 is merely an afterthought ;-)
+So in a nutshell: we greatly encourage Farmers to have IPv6 on the Node's network.
+
+### Routing/firewalling
+
+Basically, the Nodes are self-protecting, in the sense that they provide no means at all to be accessed through listening processes at all. No service is active on the node itself, and User Networks function solely on an overlay.
+That also means that there is no need for a Farm admin to protect the Nodes from exterior access, albeit some DDoS protection might be a good idea.
+In the first phase we will still allow the Host OS (ZOS) to reply on ICMP ping requests, but that 'feature' might as well be blocked in the future, as once a Node is able to register itself, there is no real need to ever want to try to reach it.
+
+### Multi-NIC Nodes
+
+Nodes that Farmers deploy are typically multi-NIC Nodes, where one (typically a 1GBit NIC) can be used for getting a proper DHCP server running from where the Nodes can boot, and one other NIC (1Gbit or even 10GBit), that then is used for transfers of User Data, so that there is a clean separation, and possible injections bogus data is not possible.
+
+That means that there would be two networks, either by different physical switches, or by port-based VLANs in the switch (if there is only one).
+
+- Management NICs
+ The Management NIC will be used by ZOS to boot, and register itself to the GRID. Also, all communications from the Node to the Grid happens from there.
+- Public NICs
+
+### Farmers and the grid
+
+A Node, being part of the Grid, has no concept of 'Farmer'. The only relationship for a Node with a Farmer is the fact that that is registered 'somewhere (TM)', and that a such workloads on a Node will be remunerated with Tokens. For the rest, a Node is a wholly stand-alone thing that participates in the Grid.
+
+```text
+ 172.16.1.0/24
+ 2a02:1807:1100:10::/64
++--------------------------------------+
+| +--------------+ | +-----------------------+
+| |Node ZOS | +-------+ | |
+| | +-------------+1GBit +--------------------+ 1GBit switch |
+| | | br-zos +-------+ | |
+| | | | | |
+| | | | | |
+| | | | +------------------+----+
+| +--------------+ | | +-----------+
+| | OOB Network | | |
+| | +----------+ ROUTER |
+| | | |
+| | | |
+| | | |
+| +------------+ | +----------+ |
+| | Public | | | | |
+| | container | | | +-----+-----+
+| | | | | |
+| | | | | |
+| +---+--------+ | +-------------------+--------+ |
+| | | | 10GBit Switch | |
+| br-pub| +-------+ | | |
+| +-----+10GBit +-------------------+ | +---------->
+| +-------+ | | Internet
+| | | |
+| | +----------------------------+
++--------------------------------------+
+ 185.69.167.128/26 Public network
+ 2a02:1807:1100:0::/64
+
+```
+
+Where the underlay part of the wireguard interfaces get instantiated in the Public container (namespace), and once created these wireguard interfaces get sent into the User Network (Network Resource), where a user can then configure the interface a he sees fit.
+
+The router of the farmer fulfills 2 roles:
+
+- NAT everything in the OOB network to the outside, so that nodes can start and register themselves, as well get tasks to execute from the BCDB.
+- Route the assigned IPv4 subnet and IPv6 public prefix on the public segment, to which the public container is connected.
+
+As such, in case that the farmer wants to provide IPv4 public access for grid proxies, the node will need at least one (1) IPv4 address. It's free to the farmer to assign IPv4 addresses to only a part of the Nodes.
+On the other hand, it is quite important to have a proper IPv6 setup, because things will work out better.
+
+It's the Farmer's task to set up the Router and the switches.
+
+In a simpler setup (small number of nodes for instance), the farmer could setup a single switch and make 2 port-based VLANs to separate OOB and Public, or even wit single-nic nodes, just put them directly on the public segment, but then he will have to provide a DHCP server on the Public network.
diff --git a/collections/documentation/developers/internals/zos/internals/network/topology/png/ndmz-dualstack.png b/collections/documentation/developers/internals/zos/internals/network/topology/png/ndmz-dualstack.png
new file mode 100644
index 0000000..dc9d508
Binary files /dev/null and b/collections/documentation/developers/internals/zos/internals/network/topology/png/ndmz-dualstack.png differ
diff --git a/collections/documentation/developers/internals/zos/internals/network/topology/png/ndmz-hidden.png b/collections/documentation/developers/internals/zos/internals/network/topology/png/ndmz-hidden.png
new file mode 100644
index 0000000..d6a462c
Binary files /dev/null and b/collections/documentation/developers/internals/zos/internals/network/topology/png/ndmz-hidden.png differ
diff --git a/collections/documentation/developers/internals/zos/internals/network/topology/png/nr-join.png b/collections/documentation/developers/internals/zos/internals/network/topology/png/nr-join.png
new file mode 100644
index 0000000..a8509b4
Binary files /dev/null and b/collections/documentation/developers/internals/zos/internals/network/topology/png/nr-join.png differ
diff --git a/collections/documentation/developers/internals/zos/internals/network/topology/png/nr-step-1.png b/collections/documentation/developers/internals/zos/internals/network/topology/png/nr-step-1.png
new file mode 100644
index 0000000..fb077d8
Binary files /dev/null and b/collections/documentation/developers/internals/zos/internals/network/topology/png/nr-step-1.png differ
diff --git a/collections/documentation/developers/internals/zos/internals/network/topology/png/nr-step-2.png b/collections/documentation/developers/internals/zos/internals/network/topology/png/nr-step-2.png
new file mode 100644
index 0000000..1d70324
Binary files /dev/null and b/collections/documentation/developers/internals/zos/internals/network/topology/png/nr-step-2.png differ
diff --git a/collections/documentation/developers/internals/zos/internals/network/topology/png/public-namespace.png b/collections/documentation/developers/internals/zos/internals/network/topology/png/public-namespace.png
new file mode 100644
index 0000000..9276012
Binary files /dev/null and b/collections/documentation/developers/internals/zos/internals/network/topology/png/public-namespace.png differ
diff --git a/collections/documentation/developers/internals/zos/internals/network/topology/png/zos-bridge.png b/collections/documentation/developers/internals/zos/internals/network/topology/png/zos-bridge.png
new file mode 100644
index 0000000..908ea0d
Binary files /dev/null and b/collections/documentation/developers/internals/zos/internals/network/topology/png/zos-bridge.png differ
diff --git a/collections/documentation/developers/internals/zos/internals/network/topology/readme.md b/collections/documentation/developers/internals/zos/internals/network/topology/readme.md
new file mode 100644
index 0000000..2ad5937
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/network/topology/readme.md
@@ -0,0 +1,68 @@
+# On boot
+> this is setup by `internet` daemon, which is part of the bootstrap process.
+
+the first basic network setup is done, the point of this setup is to connect the node to the internet, to be able to continue the rest of the boot process.
+
+- Go over all **PLUGGED, and PHYSICAL** interfaces
+- For each matching interface, the interface is tested if it can get both IPv4 and IPv6
+- If multiple interfaces have been found to receive ipv4 from dhcp, we find the `smallest` ip, with the private gateway IP, otherwise if no private gateway ip found, we only find the one with the smallest IP.
+- Once the interface is found we do the following: (we will call this interface **eth**)
+ - Create a bridge named `zos`
+ - Disable IPv6 on this bridge, and ipv6 forwarding
+- Run `udhcpc` on zos bridge
+![zos-bridge](png/zos-bridge.png)
+
+Once this setup complete, the node now has access to the internet which allows it to download and run `networkd` which takes over the network stack and continue the process as follows.
+
+# Network Daemon
+- Validate zos setup created by the `internet` on boot daemon
+- Send information about all local nics to the explorer (?)
+
+## Setting up `ndmz`
+First we need to find the master interface for ndmz, we have the following cases:
+- master of `public_config` if set. Public Config is an external configuration that is set by the farmer on the node object. that information is retrieved by the node from the public explorer.
+- otherwise (if public_config is not set) check if the public namespace is set (i think that's a dead branch because if this exist (or can exist) it means the master is always set. which means it will get used always.
+- otherwise find first interface with ipv6
+- otherwise check if zos has global unicast ipv6
+- otherwise hidden node (still uses zos but in hidden node setup)
+
+### Hidden node ndmz
+![ndmz-hidden](png/ndmz-hidden.png)
+
+### Dualstack ndmz
+![ndmz-dualstack](png/ndmz-dualstack.png)
+
+## Setting up Public Config
+this is an external configuration step that is configured by the farmer on the node object. The node then must have setup in the explorer.
+
+![public-namespace](png/public-namespace.png)
+
+## Setting up Yggdrasil
+- Get a list of all public peers with status `up`
+- If hidden node:
+ - Find peers with IPv4 addresses
+- If dual stack node:
+ - Filter out all peers with same prefix as the node, to avoid connecting locally only
+- write down yggdrasil config, and start yggdrasil daemon via zinit
+- yggdrasil runs inside the ndmz namespace
+- add an ipv6 address to npub in the same prefix as yggdrasil. this way when npub6 is used as a gateway for this prefix, traffic
+will be routed through yggdrasil.
+
+# Creating a network resource
+A network resource (`NR` for short) as a user private network that lives on the node and can span multiple nodes over wireguard. When a network is deployed the node builds a user namespace as follows:
+- A unique network id is generated by md5sum(user_id + network_name) then only take first 13 bytes. We will call this `net-id`.
+
+![nr-1](png/nr-step-1.png)
+
+## Create the wireguard interface
+if the node has `public_config` so the `public` namespace exists. then the wireguard device is first created inside the `public` namespace then moved
+to the network-resource namespace.
+
+Otherwise, the port is created on the host namespace and then moved to the network-resource namespace. The final result is
+
+![nr-2](png/nr-step-2.png)
+
+Finally the wireguard peer list is applied and configured, routing rules is also configured to route traffic to the wireguard interface
+
+# Member joining a user network (network resource)
+![nr-join](png/nr-join.png)
diff --git a/collections/documentation/developers/internals/zos/internals/network/topology/uml/ndmz-dualstack.wsd b/collections/documentation/developers/internals/zos/internals/network/topology/uml/ndmz-dualstack.wsd
new file mode 100644
index 0000000..8b3d6eb
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/network/topology/uml/ndmz-dualstack.wsd
@@ -0,0 +1,57 @@
+@startuml
+[zos\nbridge] as zos
+[br-pub\nbridge] as brpub
+[br-ndmz\nbridge] as brndmz
+note top of brndmz
+disable ipv6
+- net.ipv6.conf.br-ndmz.disable_ipv6 = 1
+end note
+' brpub -left- zos : veth pair\n(tozos)
+brpub -down- master
+note right of master
+master is found as described
+in the readme (this can be zos bridge)
+in case of a single node machine
+end note
+
+package "ndmz namespace" {
+ [tonrs\nmacvlan] as tonrs
+ note bottom of tonrs
+ - net.ipv4.conf.tonrs.proxy_arp = 0
+ - net.ipv6.conf.tonrs.disable_ipv6 = 0
+
+ Addresses:
+ 100.127.0.1/16
+ fe80::1/64
+ fd00::1
+ end note
+ tonrs - brndmz: macvlan
+
+ [npub6\nmacvlan] as npub6
+ npub6 -down- brpub: macvlan
+
+ [npub4\nmacvlan] as npub4
+ npub4 -down- zos: macvlan
+
+ note as MAC
+ gets static mac address generated
+ from node id. to make sure it receives
+ same ip address.
+ end note
+
+ MAC .. npub4
+ MAC .. npub6
+
+ note as setup
+ - net.ipv6.conf.all.forwarding = 1
+ end note
+
+ [ygg0]
+ note bottom of ygg0
+ this will be added by yggdrasil setup
+ in the next step
+ end note
+}
+
+footer (hidden node) no master with global unicast ipv6 found
+@enduml
diff --git a/collections/documentation/developers/internals/zos/internals/network/topology/uml/ndmz-hidden.wsd b/collections/documentation/developers/internals/zos/internals/network/topology/uml/ndmz-hidden.wsd
new file mode 100644
index 0000000..05304dc
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/network/topology/uml/ndmz-hidden.wsd
@@ -0,0 +1,55 @@
+@startuml
+[zos\nbridge] as zos
+note left of zos
+current select master
+for hiddent ndmz setup
+end note
+[br-pub\nbridge] as brpub
+[br-ndmz\nbridge] as brndmz
+note top of brndmz
+disable ipv6
+- net.ipv6.conf.br-ndmz.disable_ipv6 = 1
+end note
+brpub -left- zos : veth pair\n(tozos)
+
+package "ndmz namespace" {
+ [tonrs\nmacvlan] as tonrs
+ note bottom of tonrs
+ - net.ipv4.conf.tonrs.proxy_arp = 0
+ - net.ipv6.conf.tonrs.disable_ipv6 = 0
+
+ Addresses:
+ 100.127.0.1/16
+ fe80::1/64
+ fd00::1
+ end note
+ tonrs - brndmz: macvlan
+
+ [npub6\nmacvlan] as npub6
+ npub6 -right- brpub: macvlan
+
+ [npub4\nmacvlan] as npub4
+ npub4 -down- zos: macvlan
+
+ note as MAC
+ gets static mac address generated
+ from node id. to make sure it receives
+ same ip address.
+ end note
+
+ MAC .. npub4
+ MAC .. npub6
+
+ note as setup
+ - net.ipv6.conf.all.forwarding = 1
+ end note
+
+ [ygg0]
+ note bottom of ygg0
+ this will be added by yggdrasil setup
+ in the next step
+ end note
+}
+
+footer (hidden node) no master with global unicast ipv6 found
+@enduml
diff --git a/collections/documentation/developers/internals/zos/internals/network/topology/uml/nr-join.wsd b/collections/documentation/developers/internals/zos/internals/network/topology/uml/nr-join.wsd
new file mode 100644
index 0000000..0c54b18
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/network/topology/uml/nr-join.wsd
@@ -0,0 +1,23 @@
+@startuml
+
+component "br-pub" as public
+component "b-\nbridge" as bridge
+package " namespace" {
+ component eth0 as eth
+ note right of eth
+ set ip as configured in the reservation
+ it must be in the subnet assinged to n-
+ in the user resource above.
+ - set default route through n-
+ end note
+ eth .. bridge: veth
+
+ component [pub\nmacvlan] as pub
+ pub .. public
+
+ note right of pub
+ only if public ipv6 is requests
+ also gets a consistent MAC address
+ end note
+}
+@enduml
diff --git a/collections/documentation/developers/internals/zos/internals/network/topology/uml/nr-step-1.wsd b/collections/documentation/developers/internals/zos/internals/network/topology/uml/nr-step-1.wsd
new file mode 100644
index 0000000..a739c60
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/network/topology/uml/nr-step-1.wsd
@@ -0,0 +1,31 @@
+@startuml
+component [b-] as bridge
+note left of bridge
+- net.ipv6.conf.b-.disable_ipv6 = 1
+end note
+
+package "n- namespace" {
+ component [n-\nmacvlan] as nic
+ bridge .. nic: macvlan
+
+ note bottom of nic
+ - nic gets the first ip ".1" in the assigned
+ user subnet.
+ - an ipv6 driven from ipv4 that is driven from the assigned ipv4
+ - fe80::1/64
+ end note
+ component [public\nmacvlan] as public
+ note bottom of public
+ - gets an ipv4 in 100.127.0.9/16 range
+ - get an ipv6 in the fd00::/64 prefix
+ - route over 100.127.0.1
+ - route over fe80::1/64
+ end note
+ note as G
+ - net.ipv6.conf.all.forwarding = 1
+ end note
+}
+
+component [br-ndmz] as brndmz
+brndmz .. public: macvlan
+@enduml
diff --git a/collections/documentation/developers/internals/zos/internals/network/topology/uml/nr-step-2.wsd b/collections/documentation/developers/internals/zos/internals/network/topology/uml/nr-step-2.wsd
new file mode 100644
index 0000000..6cfcb68
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/network/topology/uml/nr-step-2.wsd
@@ -0,0 +1,33 @@
+@startuml
+component [b-] as bridge
+note left of bridge
+- net.ipv6.conf.b-.disable_ipv6 = 1
+end note
+
+package "n- namespace" {
+ component [n-\nmacvlan] as nic
+ bridge .. nic: macvlan
+
+ note bottom of nic
+ - nic gets the first ip ".1" in the assigned
+ user subnet.
+ - an ipv6 driven from ipv4 that is driven from the assigned ipv4
+ - fe80::1/64
+ end note
+ component [public\nmacvlan] as public
+ note bottom of public
+ - gets an ipv4 in 100.127.0.9/16 range
+ - get an ipv6 in the fd00::/64 prefix
+ - route over 100.127.0.1
+ - route over fe80::1/64
+ end note
+ note as G
+ - net.ipv6.conf.all.forwarding = 1
+ end note
+ component [w-\nwireguard]
+}
+
+
+component [br-ndmz] as brndmz
+brndmz .. public: macvlan
+@enduml
diff --git a/collections/documentation/developers/internals/zos/internals/network/topology/uml/public-namespace.wsd b/collections/documentation/developers/internals/zos/internals/network/topology/uml/public-namespace.wsd
new file mode 100644
index 0000000..215152c
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/network/topology/uml/public-namespace.wsd
@@ -0,0 +1,29 @@
+@startuml
+
+() "br-pub (Public Bridge)" as brpub
+
+note bottom of brpub
+This bridge is always created on boot, and is either
+connected to the zos bridge (in single nic setup).
+or to the seond nic with public IPv6 (in dual nic setup)
+end note
+
+
+package "public namespace" {
+
+ [public\nmacvlan] as public
+ public -down- brpub: macvlan
+ note right of public
+ - have a static mac generated from node id
+ - set the ips as configured
+ - set the default gateways as configured
+ end note
+
+ note as global
+ inside namespace
+ - net.ipv6.conf.all.accept_ra = 2
+ - net.ipv6.conf.all.accept_ra_defrtr = 1
+ end note
+}
+
+@enduml
diff --git a/collections/documentation/developers/internals/zos/internals/network/topology/uml/zos-bridge.wsd b/collections/documentation/developers/internals/zos/internals/network/topology/uml/zos-bridge.wsd
new file mode 100644
index 0000000..2328f00
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/network/topology/uml/zos-bridge.wsd
@@ -0,0 +1,16 @@
+@startuml
+() eth
+[zos]
+eth -up- zos
+note left of zos
+bridge takes same mac address as eth
+(ipv6 is enabled on the bridge)
+- net.ipv6.conf.zos.disable_ipv6 = 0
+end note
+note left of eth
+disable ipv6 on interface:
+(ipv6 is disabled on the nic)
+- net.ipv6.conf..disable_ipv6 = 1
+- net.ipv6.conf.all.forwarding = 0
+end note
+@enduml
diff --git a/collections/documentation/developers/internals/zos/internals/network/yggdrasil.md b/collections/documentation/developers/internals/zos/internals/network/yggdrasil.md
new file mode 100644
index 0000000..7b3c189
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/network/yggdrasil.md
@@ -0,0 +1,25 @@
+# Yggdrasil integration in 0-OS
+
+Since day one, 0-OS v2 networking has been design around IPv6. The goal was avoid having to deal with exhausted IPV4 address and be ready for the future.
+
+While this decision made sense on the long term, it pose trouble on the short term for farmer that only have access to ipv4 and are unable to ask for an upgrade to their IPS.
+
+In order to allow these ipv4 only nodes to join the grid, an other overlay network has to be created between all the nodes. To achieve this, Yggdrasil has been selected.
+
+## Yggdrasil
+
+[Yggdrasil network project](https://yggdrasil-network.github.io/) has been selected to be integrated into 0-OS. All 0-OS node will runs an yggdrasil daemon which means all 0-OS nodes can now communicate over the yggdrasil network. The yggdrasil integration is an experiment planned in multiple phase:
+
+Phase 1: Allow 0-DB container to be exposed over yggdrasil network. Implemented in v0.3.5
+Phase 2: Allow containers to request an interface with an yggdrasil IP address.
+
+## networkd bootstrap
+
+When booting, networkd will wait for 2 minute to receive an IPv6 address through router advertisement for it's `npub6` interface in the ndmz network namspace.
+If after 2 minutes, no IPv6 is received, networkd will consider the node to be an IPv4 only nodes, switch to this mode and continue booting.
+
+### 0-DB containers
+
+For ipv4 only nodes, the 0-DB container will be exposed on top an yggdrasil IPv6 address. Since all the 0-OS node will also run yggdrasil, these 0-DB container will always be reachable from any container in the grid.
+
+For dual stack nodes, the 0-DB container will also get an yggdrasil IP in addition to the already present public IPv6.
\ No newline at end of file
diff --git a/collections/documentation/developers/internals/zos/internals/network/zbus.md b/collections/documentation/developers/internals/zos/internals/network/zbus.md
new file mode 100644
index 0000000..c2b7a2a
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/network/zbus.md
@@ -0,0 +1,46 @@
+# Network module
+
+## ZBus
+
+Network module is available on zbus over the following channel
+
+| module | object | version |
+|--------|--------|---------|
+| network|[network](#interface)| 0.0.1|
+
+## Home Directory
+
+network keeps some data in the following locations
+| directory | path|
+|----|---|
+| root| `/var/cache/modules/network`|
+
+
+## Interface
+
+```go
+//Networker is the interface for the network module
+type Networker interface {
+ // Create a new network resource
+ CreateNR(Network) (string, error)
+ // Delete a network resource
+ DeleteNR(Network) error
+
+ // Join a network (with network id) will create a new isolated namespace
+ // that is hooked to the network bridge with a veth pair, and assign it a
+ // new IP from the network resource range. The method return the new namespace
+ // name.
+ // The member name specifies the name of the member, and must be unique
+ // The NetID is the network id to join
+ Join(networkdID NetID, containerID string, addrs []string) (join Member, err error)
+
+ // ZDBPrepare creates a network namespace with a macvlan interface into it
+ // to allow the 0-db container to be publicly accessible
+ // it retusn the name of the network namespace created
+ ZDBPrepare() (string, error)
+
+ // Addrs return the IP addresses of interface
+ // if the interface is in a network namespace netns needs to be not empty
+ Addrs(iface string, netns string) ([]net.IP, error)
+}
+```
\ No newline at end of file
diff --git a/collections/documentation/developers/internals/zos/internals/node/readme.md b/collections/documentation/developers/internals/zos/internals/node/readme.md
new file mode 100644
index 0000000..0679bd7
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/node/readme.md
@@ -0,0 +1,50 @@
+ Node Module
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Zbus](#zbus)
+- [Example](#example)
+
+***
+
+## Introduction
+
+This module is responsible of registering the node on the grid, and handling of grid events. The node daemon broadcast the intended events on zbus for other modules that are interested in those events.
+
+The node also provide zbus interfaces to query some of the node information.
+
+## Zbus
+
+Node module is available on [zbus](https://github.com/threefoldtech/zbus) over the following channel
+
+| module | object | version |
+|--------|--------|---------|
+|host |host| 0.0.1
+|system |system| 0.0.1
+|events |events| 0.0.1
+
+## Example
+
+```go
+
+//SystemMonitor interface (provided by noded)
+type SystemMonitor interface {
+ NodeID() uint32
+ Memory(ctx context.Context) <-chan VirtualMemoryStat
+ CPU(ctx context.Context) <-chan TimesStat
+ Disks(ctx context.Context) <-chan DisksIOCountersStat
+ Nics(ctx context.Context) <-chan NicsIOCounterStat
+}
+
+// HostMonitor interface (provided by noded)
+type HostMonitor interface {
+ Uptime(ctx context.Context) <-chan time.Duration
+}
+
+// Events interface
+type Events interface {
+ PublicConfigEvent(ctx context.Context) <-chan PublicConfigEvent
+ ContractCancelledEvent(ctx context.Context) <-chan ContractCancelledEvent
+}
+```
diff --git a/collections/documentation/developers/internals/zos/internals/provision/readme.md b/collections/documentation/developers/internals/zos/internals/provision/readme.md
new file mode 100644
index 0000000..c82be5c
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/provision/readme.md
@@ -0,0 +1,35 @@
+Provision Module
+
+ Table of Contents
+
+- [ZBus](#zbus)
+- [Introduction](#introduction)
+- [Supported workload](#supported-workload)
+
+
+***
+
+## ZBus
+
+This module is autonomous module and is not reachable over `zbus`.
+
+## Introduction
+
+This module is responsible to provision/decommission workload on the node.
+
+It accepts new deployment over `rmb` and tries to bring them to reality by running a series of provisioning workflows based on the workload `type`.
+
+`provisiond` knows about all available daemons and it contacts them over `zbus` to ask for the needed services. The pull everything together and update the deployment with the workload state.
+
+If node was restarted, `provisiond` tries to bring all active workloads back to original state.
+## Supported workload
+
+0-OS currently support 8 type of workloads:
+- network
+- `zmachine` (virtual machine)
+- `zmount` (disk): usable only by a `zmachine`
+- `public-ip` (v4 and/or v6): usable only by a `zmachine`
+- [`zdb`](https://github.com/threefoldtech/0-DB) `namespace`
+- [`qsfs`](https://github.com/threefoldtech/quantum-storage)
+- `zlogs`
+- `gateway`
diff --git a/collections/documentation/developers/internals/zos/internals/storage/readme.md b/collections/documentation/developers/internals/zos/internals/storage/readme.md
new file mode 100644
index 0000000..8114ec1
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/storage/readme.md
@@ -0,0 +1,153 @@
+ Storage Module
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [ZBus](#zbus)
+- [Overview](#overview)
+- [List of sub-modules](#list-of-sub-modules)
+- [On Node Booting](#on-node-booting)
+ - [zinit unit](#zinit-unit)
+ - [Interface](#interface)
+
+***
+
+## Introduction
+
+This module is responsible to manage everything related with storage.
+
+## ZBus
+
+Storage module is available on zbus over the following channel
+
+| module | object | version |
+|--------|--------|---------|
+| storage|[storage](#interface)| 0.0.1|
+
+
+## Overview
+
+On start, storaged holds ownership of all node disks, and it separate it into 2 different sets:
+
+- SSD Storage: For each ssd disk available, a storage pool of type SSD is created
+- HDD Storage: For each HDD disk available, a storage pool of type HDD is created
+
+
+Then `storaged` can provide the following storage primitives:
+- `subvolume`: (with quota). The btrfs subvolume can be used by used by `flistd` to support read-write operations on flists. Hence it can be used as rootfs for containers and VMs. This storage primitive is only supported on `ssd` pools.
+ - On boot, storaged will always create a permanent subvolume with id `zos-cache` (of 100G) which will be used by the system to persist state and to hold cache of downloaded files.
+- `vdisk`: Virtual disk that can be attached to virtual machines. this is only possible on `ssd` pools.
+- `device`: that is a full disk that gets allocated and used by a single `0-db` service. Note that a single 0-db instance can serve multiple zdb namespaces for multiple users. This is only possible for on `hdd` pools.
+
+You already can tell that ZOS can work fine with no HDD (it will not be able to server zdb workloads though), but not without SSD. Hence a zos with no SSD will never register on the grid.
+
+## List of sub-modules
+
+- disks
+- 0-db
+- booting
+
+## On Node Booting
+
+When the module boots:
+
+- Make sure to mount all available pools
+- Scan available disks that are not used by any pool and create new pools on those disks. (all pools now are created with `RaidSingle` policy)
+- Try to find and mount a cache sub-volume under /var/cache.
+- If no cache sub-volume is available a new one is created and then mounted.
+
+### zinit unit
+
+The zinit unit file of the module specify the command line, test command, and the order where the services need to be booted.
+
+Storage module is a dependency for almost all other system modules, hence it has high boot presidency (calculated on boot) by zinit based on the configuration.
+
+The storage module is only considered running, if (and only if) the /var/cache is ready
+
+```yaml
+exec: storaged
+test: mountpoint /var/cache
+```
+
+### Interface
+
+```go
+
+// StorageModule is the storage subsystem interface
+// this should allow you to work with the following types of storage medium
+// - full disks (device) (these are used by zdb)
+// - subvolumes these are used as a read-write layers for 0-fs mounts
+// - vdisks are used by zmachines
+// this works as following:
+// a storage module maintains a list of ALL disks on the system
+// separated in 2 sets of pools (SSDs, and HDDs)
+// ssd pools can only be used for
+// - subvolumes
+// - vdisks
+// hdd pools are only used by zdb as one disk
+type StorageModule interface {
+ // Cache method return information about zos cache volume
+ Cache() (Volume, error)
+
+ // Total gives the total amount of storage available for a device type
+ Total(kind DeviceType) (uint64, error)
+ // BrokenPools lists the broken storage pools that have been detected
+ BrokenPools() []BrokenPool
+ // BrokenDevices lists the broken devices that have been detected
+ BrokenDevices() []BrokenDevice
+ //Monitor returns stats stream about pools
+ Monitor(ctx context.Context) <-chan PoolsStats
+
+ // Volume management
+
+ // VolumeCreate creates a new volume
+ VolumeCreate(name string, size gridtypes.Unit) (Volume, error)
+
+ // VolumeUpdate updates the size of an existing volume
+ VolumeUpdate(name string, size gridtypes.Unit) error
+
+ // VolumeLookup return volume information for given name
+ VolumeLookup(name string) (Volume, error)
+
+ // VolumeDelete deletes a volume by name
+ VolumeDelete(name string) error
+
+ // VolumeList list all volumes
+ VolumeList() ([]Volume, error)
+
+ // Virtual disk management
+
+ // DiskCreate creates a virtual disk given name and size
+ DiskCreate(name string, size gridtypes.Unit) (VDisk, error)
+
+ // DiskResize resizes the disk to given size
+ DiskResize(name string, size gridtypes.Unit) (VDisk, error)
+
+ // DiskWrite writes the given raw image to disk
+ DiskWrite(name string, image string) error
+
+ // DiskFormat makes sure disk has filesystem, if it already formatted nothing happens
+ DiskFormat(name string) error
+
+ // DiskLookup looks up vdisk by name
+ DiskLookup(name string) (VDisk, error)
+
+ // DiskExists checks if disk exists
+ DiskExists(name string) bool
+
+ // DiskDelete deletes a disk
+ DiskDelete(name string) error
+
+ DiskList() ([]VDisk, error)
+ // Device management
+
+ //Devices list all "allocated" devices
+ Devices() ([]Device, error)
+
+ // DeviceAllocate allocates a new device (formats and give a new ID)
+ DeviceAllocate(min gridtypes.Unit) (Device, error)
+
+ // DeviceLookup inspects a previously allocated device
+ DeviceLookup(name string) (Device, error)
+}
+```
diff --git a/collections/documentation/developers/internals/zos/internals/vmd/readme.md b/collections/documentation/developers/internals/zos/internals/vmd/readme.md
new file mode 100644
index 0000000..d30fa3e
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/internals/vmd/readme.md
@@ -0,0 +1,66 @@
+VMD Module
+
+ Table of Contents
+
+- [ZBus](#zbus)
+- [Home Directory](#home-directory)
+- [Introduction](#introduction)
+ - [zinit unit](#zinit-unit)
+- [Interface](#interface)
+
+***
+
+## ZBus
+
+Storage module is available on zbus over the following channel
+
+| module | object | version |
+|--------|--------|---------|
+| vmd|[vmd](#interface)| 0.0.1|
+
+## Home Directory
+
+contd keeps some data in the following locations
+| directory | path|
+|----|---|
+| root| `/var/cache/modules/containerd`|
+
+## Introduction
+
+The vmd module, manages all virtual machines processes, it provide the interface to, create, inspect, and delete virtual machines. It also monitor the vms to make sure they are re-spawned if crashed. Internally it uses `cloud-hypervisor` to start the Vm processes.
+
+It also provide the interface to configure VM logs streamers.
+
+### zinit unit
+
+`contd` must run after containerd is running, and the node boot process is complete. Since it doesn't keep state, no dependency on `stroaged` is needed
+
+```yaml
+exec: vmd --broker unix:///var/run/redis.sock
+after:
+ - boot
+ - networkd
+```
+
+## Interface
+
+```go
+
+// VMModule defines the virtual machine module interface
+type VMModule interface {
+ Run(vm VM) error
+ Inspect(name string) (VMInfo, error)
+ Delete(name string) error
+ Exists(name string) bool
+ Logs(name string) (string, error)
+ List() ([]string, error)
+ Metrics() (MachineMetrics, error)
+
+ // VM Log streams
+
+ // StreamCreate creates a stream for vm `name`
+ StreamCreate(name string, stream Stream) error
+ // delete stream by stream id.
+ StreamDelete(id string) error
+}
+```
diff --git a/collections/documentation/developers/internals/zos/manual/api.md b/collections/documentation/developers/internals/zos/manual/api.md
new file mode 100644
index 0000000..1ffdad4
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/manual/api.md
@@ -0,0 +1,273 @@
+API
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Deployments](#deployments)
+ - [Deploy](#deploy)
+ - [Update](#update)
+ - [Get](#get)
+ - [Changes](#changes)
+ - [Delete](#delete)
+- [Statistics](#statistics)
+- [Storage](#storage)
+ - [List separate pools with capacity](#list-separate-pools-with-capacity)
+- [Network](#network)
+ - [List Wireguard Ports](#list-wireguard-ports)
+ - [Supports IPV6](#supports-ipv6)
+ - [List Public Interfaces](#list-public-interfaces)
+ - [List Public IPs](#list-public-ips)
+ - [Get Public Config](#get-public-config)
+- [Admin](#admin)
+ - [List Physical Interfaces](#list-physical-interfaces)
+ - [Get Public Exit NIC](#get-public-exit-nic)
+ - [Set Public Exit NIC](#set-public-exit-nic)
+- [System](#system)
+ - [Version](#version)
+ - [DMI](#dmi)
+ - [Hypervisor](#hypervisor)
+- [GPUs](#gpus)
+ - [List Gpus](#list-gpus)
+
+
+***
+
+## Introduction
+
+This document should list all the actions available on the node public API. which is available over [RMB](https://github.com/threefoldtech/rmb-rs)
+
+The node is always reachable over the node twin id as per the node object on tfchain. Once node twin is known, a [client](https://github.com/threefoldtech/zos/blob/main/client/node.go) can be initiated and used to talk to the node.
+
+## Deployments
+
+### Deploy
+
+| command |body| return|
+|---|---|---|
+| `zos.deployment.deploy` | [Deployment](https://github.com/threefoldtech/zos/blob/main/pkg/gridtypes/deployment.go)|-|
+
+Deployment need to have valid signature, the contract must exist on chain with the correct contract hash as the deployment.
+
+### Update
+
+| command |body| return|
+|---|---|---|
+| `zos.deployment.update` | [Deployment](https://github.com/threefoldtech/zos/blob/main/pkg/gridtypes/deployment.go)|-|
+
+The update call, will update (modify) an already existing deployment with new definition. The deployment must already exist on the node, the contract must have the new hash as the provided deployment, plus valid versions.
+
+> TODO: need more details over the deployment update calls how to handle the version
+
+### Get
+
+| command |body| return|
+|---|---|---|
+| `zos.deployment.get` | `{contract_id: }`|[Deployment](https://github.com/threefoldtech/zos/blob/main/pkg/gridtypes/deployment.go)|
+
+### Changes
+
+| command |body| return|
+|---|---|---|
+| `zos.deployment.changes` | `{contract_id: }`| `[]Workloads` |
+
+Where:
+
+- [workload](https://github.com/threefoldtech/zos/blob/main/pkg/gridtypes/workload.go)
+
+The list will contain all deployment workloads (changes) means a workload can (will) appear
+multiple times in this list for each time a workload state will change.
+
+This means a workload will first appear in `init` state, then next time it will show the state change (with time) to the next state which can be success or failure, and so on.
+This will happen for each workload in the deployment.
+
+### Delete
+>
+> You probably never need to call this command yourself, the node will delete the deployment once the contract is cancelled on the chain.
+
+| command |body| return|
+|---|---|---|
+| `zos.deployment.get` | `{contract_id: }`|-|
+
+## Statistics
+
+| command |body| return|
+|---|---|---|
+| `zos.statistics.get` | - |`{total: Capacity, used: Capacity, system: Capacity}`|
+
+Where:
+
+```json
+Capacity {
+ "cur": "uint64",
+ "sru": "bytes",
+ "hru": "bytes",
+ "mru": "bytes",
+ "ipv4u": "unit64",
+}
+```
+
+> Note that, `used` capacity equal the full workload reserved capacity PLUS the system reserved capacity
+so `used = user_used + system`, while `system` is only the amount of resourced reserved by `zos` itself
+
+## Storage
+
+### List separate pools with capacity
+
+| command |body| return|
+|---|---|---|
+| `zos.storage.pools` | - |`[]Pool`|
+
+List all node pools with their types, size and used space
+where
+
+```json
+Pool {
+ "name": "pool-id",
+ "type": "(ssd|hdd)",
+ "size": ,
+ "used":
+}
+```
+
+## Network
+
+### List Wireguard Ports
+
+| command |body| return|
+|---|---|---|
+| `zos.network.list_wg_ports` | - |`[]uint16`|
+
+List all `reserved` ports on the node that can't be used for network wireguard. A user then need to find a free port that is not in this list to use for his network
+
+### Supports IPV6
+
+| command |body| return|
+|---|---|---|
+| `zos.network.has_ipv6` | - |`bool`|
+
+### List Public Interfaces
+
+| command |body| return|
+|---|---|---|
+| `zos.network.interfaces` | - |`map[string][]IP` |
+
+list of node IPs this is a public information. Mainly to show the node yggdrasil IP and the `zos` interface.
+
+### List Public IPs
+
+| command |body| return|
+|---|---|---|
+| `zos.network.list_public_ips` | - |`[]IP` |
+
+List all user deployed public IPs that are served by this node.
+
+### Get Public Config
+
+| command |body| return|
+|---|---|---|
+| `zos.network.public_config_get` | - |`PublicConfig` |
+
+Where
+
+```json
+PublicConfig {
+ "type": "string", // always vlan
+ "ipv4": "CIDR",
+ "ipv6": "CIDR",
+ "gw4": "IP",
+ "gw6": "IP",
+ "domain": "string",
+}
+```
+
+returns the node public config or error if not set. If a node has public config
+it means it can act like an access node to user private networks
+
+## Admin
+
+The next set of commands are ONLY possible to be called by the `farmer` only.
+
+### List Physical Interfaces
+
+| command |body| return|
+|---|---|---|
+| `zos.network.admin.interfaces` | - |`map[string]Interface` |
+
+Where
+
+```json
+Interface {
+ "ips": ["ip"],
+ "mac": "mac-address",
+}
+```
+
+Lists ALL node physical interfaces.
+Those interfaces then can be used as an input to `set_public_nic`
+
+### Get Public Exit NIC
+
+| command |body| return|
+|---|---|---|
+| `zos.network.admin.get_public_nic` | - |`ExitDevice` |
+
+Where
+
+```json
+ExitInterface {
+ "is_single": "bool",
+ "is_dual": "bool",
+ "dual_interface": "name",
+}
+```
+
+returns the interface used by public traffic (for user workloads)
+
+### Set Public Exit NIC
+
+| command |body| return|
+|---|---|---|
+| `zos.network.admin.set_public_nic` | `name` |- |
+
+name must be one of (free) names returned by `zos.network.admin.interfaces`
+
+## System
+
+### Version
+
+| command |body| return|
+|---|---|---|
+| `zos.system.version` | - | `{zos: string, zinit: string}` |
+
+### DMI
+
+| command |body| return|
+|---|---|---|
+| `zos.system.dmi` | - | [DMI](https://github.com/threefoldtech/zos/blob/main/pkg/capacity/dmi/dmi.go) |
+
+### Hypervisor
+
+| command |body| return|
+|---|---|---|
+| `zos.system.hypervisor` | - | `string` |
+
+## GPUs
+
+### List Gpus
+
+| command |body| return|
+|---|---|---|
+| `zos.gpu.list` | - | `[]GPU` |
+
+Where
+
+```json
+GPU {
+ "id": "string"
+ "vendor": "string"
+ "device": "string",
+ "contract": "uint64",
+}
+```
+
+Lists all available node GPUs if exist
diff --git a/collections/documentation/developers/internals/zos/manual/gateway/fqdn-proxy.md b/collections/documentation/developers/internals/zos/manual/gateway/fqdn-proxy.md
new file mode 100644
index 0000000..07a2f8b
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/manual/gateway/fqdn-proxy.md
@@ -0,0 +1,5 @@
+# `gateway-fqdn-proxy` type
+
+This create a proxy with the given fqdn to the given backends. In this case the user then must configure his dns server (i.e name.com) to point to the correct node public IP.
+
+Full name-proxy workload data is defined [here](https://github.com/threefoldtech/zos/blob/main/pkg/gridtypes/zos/gw_fqdn.go)
diff --git a/collections/documentation/developers/internals/zos/manual/gateway/name-proxy.md b/collections/documentation/developers/internals/zos/manual/gateway/name-proxy.md
new file mode 100644
index 0000000..2ce40f3
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/manual/gateway/name-proxy.md
@@ -0,0 +1,5 @@
+# `gateway-name-proxy` type
+
+This create a proxy with the given name to the given backends. The `name` of the proxy must be owned by a name contract on the grid. The idea is that a user can reserve a name (i.e `example`). Later he can deploy a gateway work load with name `example` on any gateway node that points to specified backends. The name then is prefix by the gateway name. For example if the gateway domain is `gent0.freefarm.com` then your full QFDN is goint to be called `example.gen0.freefarm.com`
+
+Full name-proxy workload data is defined [here](https://github.com/threefoldtech/zos/blob/main/pkg/gridtypes/zos/gw_name.go)
diff --git a/collections/documentation/developers/internals/zos/manual/ip/readme.md b/collections/documentation/developers/internals/zos/manual/ip/readme.md
new file mode 100644
index 0000000..0e6068f
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/manual/ip/readme.md
@@ -0,0 +1,11 @@
+# `ip` type
+The IP workload type reserves an IP from the available contract IPs list. Which means on contract creation the user must specify number of public IPs it needs to use. The contract then will allocate this number of IPs from the farm and will kept on the contract.
+
+When the user then add the IP workload to the deployment associated with this contract, each IP workload will pick and link to one IP from the contract.
+
+In minimal form, `IP` workload does not require any data. But in reality it has 2 flags to pick which kind of public IP do you want
+
+- `ipv4` (`bool`): pick one from the contract public Ipv4
+- `ipv6` (`bool`): pick an IPv6 over SLAAC. Ipv6 are not reserved with a contract. They are basically free if the farm infrastructure allows Ipv6 over SLAAC.
+
+Full `IP` workload definition can be found [here](https://github.com/threefoldtech/zos/blob/main/pkg/gridtypes/zos/ipv4.go)
diff --git a/collections/documentation/developers/internals/zos/manual/manual.md b/collections/documentation/developers/internals/zos/manual/manual.md
new file mode 100644
index 0000000..fb926c9
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/manual/manual.md
@@ -0,0 +1,187 @@
+ ZOS Manual
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Farm? Network? What are these?](#farm-network-what-are-these)
+- [Creating a farm](#creating-a-farm)
+- [Interaction](#interaction)
+- [Deployment](#deployment)
+ - [Workload](#workload)
+ - [Types](#types)
+ - [API](#api)
+- [Raid Controller Configuration](#raid-controller-configuration)
+
+***
+
+## Introduction
+
+This document explain the usage of `ZOS`. `ZOS` usually pronounced (zero OS), got it's name from the idea of zero configuration. Since after the initial `minimal` configuration which only include which `farm` to join and what `network` (`development`, `testing`, or `production`) the owner of the node does not has to do anything more, and the node work fully autonomous.
+
+The farmer himself cannot control the node, or access it by any mean. The only way you can interact with a node is via it's public API.
+
+## Farm? Network? What are these?
+
+Well, `zos` is built to allow people to run `workloads` around the world this simply is enabled by allowing 3rd party data-centers to run `ZOS` on their hardware. Then a user can then find any nearby `farm` (is what we call a cluster of nodes that belong to the same `farmer`) and then they can choose to deploy capacity on that node/farm. A `farm` can consist of one or more nodes.
+
+So what is `network`.Well, to allow developers to build and `zos` itself and make it available during the early stages of development for testers and other enthusiastic people to try it out. To allow this we created 3 `networks`
+- `development`: This is used mainly by developers to test their work. This is still available for users to deploy their capacity on (for really really cheap prices), but at the same time there is no grantee that it's stable or that data loss or corruption will happen. Also the entire network can be reset with no heads up.
+- `testing`: Once new features are developed and well tested on `development` network they are released to `testing` environment. This also available for users to use with a slightly higher price than `development` network. But it's much more stable. In theory this network is stable, there should be no resets of the network, issues on this network usually are not fatal, but partial data loss can still occurs.
+- `production`: Well, as the name indicates this is the most stable network (also full price) once new features are fully tested on `testing` network they are released on `production`.
+
+## Creating a farm
+
+While this is outside the scope of this document here you are a [link](https://library.threefold.me/info/manual/#/manual__create_farm)
+
+## Interaction
+
+`ZOS` provide a simple `API` that can be used to:
+- Query node runtime information
+ - Network information
+ - Free `wireguard` ports
+ - Get public configuration
+ - System version
+ - Other (check client for details)
+- Deployment management (more on that later)
+ - Create
+ - Update
+ - Delete
+
+Note that `zos` API is available over `rmb` protocol. `rmb` which means `reliable message bus` is a simple messaging protocol that enables peer to peer communication over `yggdrasil` network. Please check [`rmb`](https://github.com/threefoldtech/rmb) for more information.
+
+Simply put, `RMB` allows 2 entities two communicate securely knowing only their `id` an id is linked to a public key on the blockchain. Hence messages are verifiable via a signature.
+
+To be able to contact the node directly you need to run
+- `yggdrasil`
+- `rmb` (correctly configured)
+
+Once you have those running you can now contact the node over `rmb`. For a reference implementation (function names and parameters) please refer to [RMB documentation](../../rmb/rmb_toc.md)
+
+Here is a rough example of how low level creation of a deployment is done.
+
+```go
+cl, err := rmb.Default()
+if err != nil {
+ panic(err)
+}
+```
+then create an instance of the node client
+```go
+node := client.NewNodeClient(NodeTwinID, cl)
+```
+define your deployment object
+```go
+dl := gridtypes.Deployment{
+ Version: Version,
+ TwinID: Twin, //LocalTwin,
+ // this contract id must match the one on substrate
+ Workloads: []gridtypes.Workload{
+ network(), // network workload definition
+ zmount(), // zmount workload definition
+ publicip(), // public ip definition
+ zmachine(), // zmachine definition
+ },
+ SignatureRequirement: gridtypes.SignatureRequirement{
+ WeightRequired: 1,
+ Requests: []gridtypes.SignatureRequest{
+ {
+ TwinID: Twin,
+ Weight: 1,
+ },
+ },
+ },
+}
+```
+compute hash
+```go
+hash, err := dl.ChallengeHash()
+if err != nil {
+ panic("failed to create hash")
+}
+fmt.Printf("Hash: %x\n", hash)
+```
+create the contract on `substrate` and get the `contract id` then you can link the deployment to the contract, then send to the node.
+
+```go
+dl.ContractID = 11 // from substrate
+ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
+defer cancel()
+err = node.DeploymentDeploy(ctx, dl)
+if err != nil {
+ panic(err)
+}
+```
+
+Once the node receives the deployment. It will then fetch the contract (using the contract id) from the node recompute the deployment hash and compare with the one set on the contract. If matches, the node proceeds to process the deployment.
+
+## Deployment
+
+A deployment is a set of workloads that are contextually related. Workloads in the same deployment can reference to other workloads in the same deployment. But can't be referenced from another deployment. Well, except the network workload which can be referenced from a different deployment as long it belongs to the same user.
+
+Workloads has unique IDs (per deployment) that are set by the user, hence he can create multiple workloads then reference to them with the given IDs (`names`)
+
+For example, a deployment can define
+- A private network with id `net`
+- A disk with id `data`
+- A public IP with id `ip`
+- A container that uses:
+ - The container can mount the disk like `mount: {data: /mount/path}`.
+ - The container can get assign the public IP to itself like by referencing the IP with id `ip`.
+ - etc.
+
+### Workload
+Each workload has a type which is associated with some data. So minimal definition of a workload contains:
+- `name`: unique per deployment (id)
+- `type`: workload type
+- `data`: workload data that is proper for the selected type.
+
+```go
+
+// Workload struct
+type Workload struct {
+ // Version is version of reservation object. On deployment creation, version must be 0
+ // then only workloads that need to be updated must match the version of the deployment object.
+ // if a deployment update message is sent to a node it does the following:
+ // - validate deployment version
+ // - check workloads list, if a version is not matching the new deployment version, the workload is untouched
+ // - if a workload version is same as deployment, the workload is "updated"
+ // - if a workload is removed, the workload is deleted.
+ Version uint32 `json:"version"`
+ //Name is unique workload name per deployment (required)
+ Name Name `json:"name"`
+ // Type of the reservation (container, zdb, vm, etc...)
+ Type WorkloadType `json:"type"`
+ // Data is the reservation type arguments.
+ Data json.RawMessage `json:"data"`
+ // Metadata is user specific meta attached to deployment, can be used to link this
+ // deployment to other external systems for automation
+ Metadata string `json:"metadata"`
+ //Description human readale description of the workload
+ Description string `json:"description"`
+ // Result of reservation, set by the node
+ Result Result `json:"result"`
+}
+```
+
+### Types
+- Virtual machine related
+ - [`network`](./workload_types.md#network-type)
+ - [`ip`](./workload_types.md#ip-type)
+ - [`zmount`](./workload_types.md#zmount-type)
+ - [`zmachine`](./workload_types.md#zmachine-type)
+ - [`zlogs`](./workload_types.md#zlogs-type)
+- Storage related
+ - [`zdb`](./workload_types.md#zdb-type)
+ - [`qsfs`](./workload_types.md#qsfs-type)
+- Gateway related
+ - [`gateway-name-proxy`](./workload_types.md#gateway-name-proxy-type)
+ - [`gateway-fqdn-proxy`](./workload_types.md#gateway-fqdn-proxy-type)
+
+### API
+Node is always connected to the RMB network with the node `twin`. Means the node is always reachable over RMB with the node `twin-id` as an address.
+
+The [node client](https://github.com/threefoldtech/zos/blob/main/client/node.go) should have a complete list of all available functions. documentations of the API can be found [here](./api.md)
+
+## Raid Controller Configuration
+
+0-OS goal is to expose raw capacity. So it is best to always try to give it access to the most raw access to the disks. In case of raid controllers, the best is to try to set it up in [JBOD](https://en.wikipedia.org/wiki/Non-RAID_drive_architectures#JBOD) mode if available.
\ No newline at end of file
diff --git a/collections/documentation/developers/internals/zos/manual/network/readme.md b/collections/documentation/developers/internals/zos/manual/network/readme.md
new file mode 100644
index 0000000..3cbc23b
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/manual/network/readme.md
@@ -0,0 +1,14 @@
+# `network` type
+Private network can span multiple nodes at the same time. Which means workloads (`VMs`) that live (on different node) but part of the same virtual network can still reach each other over this `private` network.
+
+If one (or more) nodes are `public access nodes` you can also add your personal laptop to the nodes and be able to reach your `VMs` over `wireguard` network.
+
+In the simplest form a network workload consists of:
+- network range
+- sub-range available on this node
+- private key
+- list of peers
+ - each peer has public key
+ - sub-range
+
+Full network definition can be found [here](https://github.com/threefoldtech/zos/blob/main/pkg/gridtypes/zos/network.go)
diff --git a/collections/documentation/developers/internals/zos/manual/qsfs/readme.md b/collections/documentation/developers/internals/zos/manual/qsfs/readme.md
new file mode 100644
index 0000000..1049cdf
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/manual/qsfs/readme.md
@@ -0,0 +1,5 @@
+# `qsfs` type
+
+`qsfs` short for `quantum safe file system` is a FUSE filesystem which aim to be able to support unlimited local storage with remote backend for offload and backup which cannot be broke even by a quantum computer. Please read about it [here](https://github.com/threefoldtech/quantum-storage)
+
+To create a `qsfs` workload you need to provide the workload type as [here](https://github.com/threefoldtech/zos/blob/main/pkg/qsfsd/qsfs.go)
diff --git a/collections/documentation/developers/internals/zos/manual/workload_types.md b/collections/documentation/developers/internals/zos/manual/workload_types.md
new file mode 100644
index 0000000..a9e2d85
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/manual/workload_types.md
@@ -0,0 +1,108 @@
+ Workload Types
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Virtual Machine](#virtual-machine)
+ - [`network` type](#network-type)
+ - [`ip` type](#ip-type)
+ - [`zmount` type](#zmount-type)
+ - [`zmachine` type](#zmachine-type)
+ - [Building your `flist`](#building-your-flist)
+ - [`zlogs` type](#zlogs-type)
+- [Storage](#storage)
+ - [`zdb` type](#zdb-type)
+ - [`qsfs` type](#qsfs-type)
+- [Gateway](#gateway)
+ - [`gateway-name-proxy` type](#gateway-name-proxy-type)
+ - [`gateway-fqdn-proxy` type](#gateway-fqdn-proxy-type)
+
+## Introduction
+
+Each workload has a type which is associated with some data. We present here the different types of workload associated with Zero-OS.
+
+## Virtual Machine
+
+### `network` type
+Private network can span multiple nodes at the same time. Which means workloads (`VMs`) that live (on different node) but part of the same virtual network can still reach each other over this `private` network.
+
+If one (or more) nodes are `public access nodes` you can also add your personal laptop to the nodes and be able to reach your `VMs` over `wireguard` network.
+
+In the simplest form a network workload consists of:
+- network range
+- sub-range available on this node
+- private key
+- list of peers
+ - each peer has public key
+ - sub-range
+
+Full network definition can be found [here](https://github.com/threefoldtech/zos/blob/main/pkg/gridtypes/zos/network.go)
+
+### `ip` type
+The IP workload type reserves an IP from the available contract IPs list. Which means on contract creation the user must specify number of public IPs it needs to use. The contract then will allocate this number of IPs from the farm and will kept on the contract.
+
+When the user then add the IP workload to the deployment associated with this contract, each IP workload will pick and link to one IP from the contract.
+
+In minimal form, `IP` workload does not require any data. But in reality it has 2 flags to pick which kind of public IP do you want
+
+- `ipv4` (`bool`): pick one from the contract public Ipv4
+- `ipv6` (`bool`): pick an IPv6 over SLAAC. Ipv6 are not reserved with a contract. They are basically free if the farm infrastructure allows Ipv6 over SLAAC.
+
+Full `IP` workload definition can be found [here](https://github.com/threefoldtech/zos/blob/main/pkg/gridtypes/zos/ipv4.go)
+
+### `zmount` type
+A `zmount` is a local disk that can be attached directly to a container or a virtual machine. `zmount` only require `size` as input as defined [here](https://github.com/threefoldtech/zos/blob/main/pkg/gridtypes/zos/zmount.go) this workload type is only utilized via the `zmachine` workload.
+
+### `zmachine` type
+
+`zmachine` is a unified container/virtual machine type. This can be used to start a virtual machine on a `zos` node give the following:
+- `flist`, this what provide the base `vm` image or container image.
+ - the `flist` content is what changes the `zmachine` mode. An `flist` built from a docker image or has files, or executable binaries will run in a container mode. `ZOS` will inject it's own `kernel+initramfs` to run the workload and kick start the defined `flist` `entrypoint`
+- private network to join (with assigned IP)
+- optional public `ipv4` or `ipv6`
+- optional disks. But at least one disk is required in case running `zmachine` in `vm` mode, which is used to hold the `vm` root image.
+
+For more details on all parameters needed to run a `zmachine` please refer to [`zmachine` data](https://github.com/threefoldtech/zos/blob/main/pkg/gridtypes/zos/zmachine.go)
+
+#### Building your `flist`
+
+Please refer to [this document](./manual.md) here about how to build an compatible `zmachine flist`
+
+### `zlogs` type
+
+Zlogs is a utility workload that allows you to stream `zmachine` logs to a remote location.
+
+The `zlogs` workload needs to know what `zmachine` to stream logs of and also the `target` location to stream the logs to. `zlogs` uses internally the [`tailstream`](https://github.com/threefoldtech/tailstream) so it supports any streaming url that is supported by this utility.
+
+`zlogs` workload runs inside the same private network as the `zmachine` instance. Which means zlogs can stream logs to other `zmachines` that is running inside the same private network (possibly on different nodes).
+
+For example, you can run [`logagg`](https://github.com/threefoldtech/logagg) which is a web-socket server that can work with `tailstream` web-socket protocol.
+
+Check `zlogs` configuration [here](https://github.com/threefoldtech/zos/blob/main/pkg/gridtypes/zos/zlogs.go)
+
+## Storage
+
+### `zdb` type
+`zdb` is a storage primitives that gives you a persisted key value store over RESP protocol. Please check [`zdb` docs](https://github.com/threefoldtech/0-db)
+
+Please check [here](https://github.com/threefoldtech/zos/blob/main/pkg/zdb/zdb.go) for workload data.
+
+### `qsfs` type
+
+`qsfs` short for `quantum safe file system` is a FUSE filesystem which aim to be able to support unlimited local storage with remote backend for offload and backup which cannot be broke even by a quantum computer. Please read about it [here](https://github.com/threefoldtech/quantum-storage)
+
+To create a `qsfs` workload you need to provide the workload type as [here](https://github.com/threefoldtech/zos/blob/main/pkg/qsfsd/qsfs.go)
+
+## Gateway
+
+### `gateway-name-proxy` type
+
+This create a proxy with the given name to the given backends. The `name` of the proxy must be owned by a name contract on the grid. The idea is that a user can reserve a name (i.e `example`). Later he can deploy a gateway work load with name `example` on any gateway node that points to specified backends. The name then is prefix by the gateway name. For example if the gateway domain is `gent0.freefarm.com` then your full QFDN is goint to be called `example.gen0.freefarm.com`
+
+Full name-proxy workload data is defined [here](https://github.com/threefoldtech/zos/blob/main/pkg/gridtypes/zos/gw_name.go)
+
+### `gateway-fqdn-proxy` type
+
+This create a proxy with the given fqdn to the given backends. In this case the user then must configure his dns server (i.e name.com) to point to the correct node public IP.
+
+Full name-proxy workload data is defined [here](https://github.com/threefoldtech/zos/blob/main/pkg/gridtypes/zos/gw_fqdn.go)
diff --git a/collections/documentation/developers/internals/zos/manual/zdb/readme.md b/collections/documentation/developers/internals/zos/manual/zdb/readme.md
new file mode 100644
index 0000000..45f88f4
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/manual/zdb/readme.md
@@ -0,0 +1,4 @@
+# `zdb` type
+`zdb` is a storage primitives that gives you a persisted key value store over RESP protocol. Please check [`zdb` docs](https://github.com/threefoldtech/0-db)
+
+Please check [here](https://github.com/threefoldtech/zos/blob/main/pkg/zdb/zdb.go) for workload data.
diff --git a/collections/documentation/developers/internals/zos/manual/zlogs/readme.md b/collections/documentation/developers/internals/zos/manual/zlogs/readme.md
new file mode 100644
index 0000000..b77ade4
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/manual/zlogs/readme.md
@@ -0,0 +1,11 @@
+# `zlogs` type
+
+Zlogs is a utility workload that allows you to stream `zmachine` logs to a remote location.
+
+The `zlogs` workload needs to know what `zmachine` to stream logs of and also the `target` location to stream the logs to. `zlogs` uses internally the [`tailstream`](https://github.com/threefoldtech/tailstream) so it supports any streaming url that is supported by this utility.
+
+`zlogs` workload runs inside the same private network as the `zmachine` instance. Which means zlogs can stream logs to other `zmachines` that is running inside the same private network (possibly on different nodes).
+
+For example, you can run [`logagg`](https://github.com/threefoldtech/logagg) which is a web-socket server that can work with `tailstream` web-socket protocol.
+
+Check `zlogs` configuration [here](https://github.com/threefoldtech/zos/blob/main/pkg/gridtypes/zos/zlogs.go)
diff --git a/collections/documentation/developers/internals/zos/manual/zmachine/cloud-console.md b/collections/documentation/developers/internals/zos/manual/zmachine/cloud-console.md
new file mode 100644
index 0000000..f1e0324
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/manual/zmachine/cloud-console.md
@@ -0,0 +1,14 @@
+# Cloud console
+
+- `cloud-console` is a tool to view machine logging and interact with the machine you have deployed
+- It always runs on the machine's private network ip and port number equla to `20000 +last octect` of machine private IP
+- For example if the machine ip is `10.20.2.2/24` this means
+ - `cloud-console` is running on `10.20.2.1:20002`
+- For the cloud-console to run we need to start the cloud-hypervisor with option "--serial pty" instead of tty, this allows us to interact with the vm from another process `cloud-console` in our case
+- To be able to connect to the web console you should first start wireguard to connect to the private network
+
+```
+wg-quick up wireguard.conf
+```
+
+- Then go to your browser with the network router IP `10.20.2.1:20002`
diff --git a/collections/documentation/developers/internals/zos/manual/zmachine/readme.md b/collections/documentation/developers/internals/zos/manual/zmachine/readme.md
new file mode 100644
index 0000000..e94a0e7
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/manual/zmachine/readme.md
@@ -0,0 +1,13 @@
+# `zmachine` type
+
+`zmachine` is a unified container/virtual machine type. This can be used to start a virtual machine on a `zos` node give the following:
+- `flist`, this what provide the base `vm` image or container image.
+ - the `flist` content is what changes the `zmachine` mode. An `flist` built from a docker image or has files, or executable binaries will run in a container mode. `ZOS` will inject it's own `kernel+initramfs` to run the workload and kick start the defined `flist` `entrypoint`
+- private network to join (with assigned IP)
+- optional public `ipv4` or `ipv6`
+- optional disks. But at least one disk is required in case running `zmachine` in `vm` mode, which is used to hold the `vm` root image.
+
+For more details on all parameters needed to run a `zmachine` please refer to [`zmachine` data](https://github.com/threefoldtech/zos/blob/main/pkg/gridtypes/zos/zmachine.go)
+
+# Building your `flist`.
+Please refer to [this document](../manual.md) here about how to build an compatible `zmachine flist`
diff --git a/collections/documentation/developers/internals/zos/manual/zmachine/zmachine.md b/collections/documentation/developers/internals/zos/manual/zmachine/zmachine.md
new file mode 100644
index 0000000..ddbd102
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/manual/zmachine/zmachine.md
@@ -0,0 +1,410 @@
+# Zmachine
+
+A `Zmachine` is an instance of virtual compute capacity. There are 2 kinds of Zmachines.
+One is a `VM`, standard in cloud environments. Next to this it can also be a `container`.
+On the Zos level, both of these are implemented as virtual machines. Depending on
+the context, it will be considered to be either a VM or a container. In either
+scenario, the `Zmachine` is started from an `Flist`.
+
+> Note, both VM and Container on ZOS are actually served as Virtual Machines. The
+only difference is that if you are running in VM mode, you only need to provide a **raw**
+disk image (image.raw) in your flist.
+
+## Container
+
+A container is meant to host `microservice`. The `microservice` architecture generally
+dictates that each service should be run in it's own container (therefore providing
+a level of isolation), and communicate with other containers it depends on over the
+network.
+
+Similar to docker. In Zos, a container is actually also run in a virtualized environment.
+Similar to containers, some setup is done on behalf of the user. After setup this is done,
+the users `entrypoint` is started.
+
+It should be noted that a container has no control over the kernel
+used to run it, if this is required, a `VM` should be used instead. Furthermore,
+a container should ideally only have 1 process running. A container can be a single
+binary, or a complete filesystem. In general, the first should be preferred, and
+if you need the latter, it might be an indication that you actually want a `VM`.
+
+For containers, the network setup will be created for you. Your init process can
+assume that it will be fully set up (according to the config you provided) by the
+time it is started. Mountpoints will also be setup for you. The environment variables
+passed will be available inside the container.
+
+## VM
+
+In container mode, zos provide a minimal kernel that is used to run a light weight VM
+and then run your app from your flist. If you need control over the kernel you can actually
+still provide it inside the flist as follows:
+
+- /boot/vmlinuz
+- /boot/initrd.img [optional]
+
+**NOTE**: the vmlinuz MUST be an EFI kernel (not compressed) if building your own kernel, or you can use the [extract-vmlinux](https://github.com/torvalds/linux/blob/master/scripts/extract-vmlinux) script to extract the EFI kernel. To test if your kernel is a valid elf kernel run command
+`readelf -n `
+
+Any of those files can be a symlink to another file in the flist.
+
+If ZOS found the `/boot/vmlinuz` file, it will use this with the initrd.img if also exists. otherwise zos will use the built-in minimal kernel and run in `container` mode.
+
+### Building an ubuntu VM flist
+
+This is a guide to help you build a working VM flist.
+
+This guide is for ubuntu `jammy`
+
+prepare rootfs
+
+```bash
+mkdir ubuntu:jammy
+```
+
+bootstrap ubuntu
+
+```bash
+sudo debootstrap jammy ubuntu:jammy http://archive.ubuntu.com/ubuntu
+```
+
+this will create and download the basic rootfs for ubuntu jammy in the directory `ubuntu:jammy`.
+After its done we can then chroot into this directory to continue installing the necessary packages needed and configure
+few things.
+
+> I am using script called `arch-chroot` which is available by default on arch but you can also install on ubuntu to continue
+the following steps
+
+```bash
+sudo arch-chroot ubuntu:jammy
+```
+
+> This script (similar to the `chroot` command) switch root to that given directory but also takes care of mounting /dev /sys, etc.. for you
+and clean it up on exit.
+
+Next just remove this link and re-create the file with a valid name to be able to continue
+
+```bash
+# make sure to set the path correctly
+export PATH=/usr/local/sbin/:/usr/local/bin/:/usr/sbin/:/usr/bin/:/sbin:/bin
+
+rm /etc/resolv.conf
+echo 'nameserver 1.1.1.1' > /etc/resolv.conf
+```
+
+Install cloud-init
+
+```bash
+apt-get update
+apt-get install cloud-init openssh-server curl
+# to make sure we have clean setup
+cloud-init clean
+```
+
+Also really important that we install a kernel
+
+```bash
+apt-get install linux-modules-extra-5.15.0-25-generic
+```
+
+> I choose this package because it will also install extra modules for us and a generic kernel
+
+Next make sure that virtiofs is part of the initramfs image
+
+```bash
+echo 'fs-virtiofs' >> /etc/initramfs-tools/modules
+update-initramfs -c -k all
+```
+
+clean up cache
+
+```bash
+apt-get clean
+```
+
+Last thing we do inside the container before we actually upload the flist
+is to make sure the kernel is in the correct format
+
+This step does not require that we stay in the chroot so hit `ctr+d` or type `exit`
+
+you should be out of the arch-chroot now
+
+```bash
+curl -O https://raw.githubusercontent.com/torvalds/linux/master/scripts/extract-vmlinux
+chmod +x extract-vmlinux
+
+sudo ./extract-vmlinux ubuntu:jammy/boot/vmlinuz | sudo tee ubuntu:jammy/boot/vmlinuz-5.15.0-25-generic.elf > /dev/null
+# then replace original kernel
+sudo mv ubuntu:jammy/boot/vmlinuz-5.15.0-25-generic.elf ubuntu:jammy/boot/vmlinuz-5.15.0-25-generic
+```
+
+To verify you can do this:
+
+```bash
+ls -l ubuntu:jammy/boot
+```
+
+and it should show something like
+
+```bash
+total 101476
+-rw-r--r-- 1 root root 260489 Mar 30 2022 config-5.15.0-25-generic
+drwxr-xr-x 1 root root 54 Jun 28 15:35 grub
+lrwxrwxrwx 1 root root 28 Jun 28 15:35 initrd.img -> initrd.img-5.15.0-25-generic
+-rw-r--r-- 1 root root 41392462 Jun 28 15:39 initrd.img-5.15.0-25-generic
+lrwxrwxrwx 1 root root 28 Jun 28 15:35 initrd.img.old -> initrd.img-5.15.0-25-generic
+-rw------- 1 root root 6246119 Mar 30 2022 System.map-5.15.0-25-generic
+lrwxrwxrwx 1 root root 25 Jun 28 15:35 vmlinuz -> vmlinuz-5.15.0-25-generic
+-rw-r--r-- 1 root root 55988436 Jun 28 15:50 vmlinuz-5.15.0-25-generic
+lrwxrwxrwx 1 root root 25 Jun 28 15:35 vmlinuz.old -> vmlinuz-5.15.0-25-generic
+```
+
+Now package the tar for upload
+
+```bash
+sudo rm -rf ubuntu:jammy/dev/*
+sudo tar -czf ubuntu-jammy.tar.gz -C ubuntu:jammy .
+```
+
+Upload to the hub, and use it to create a Zmachine
+
+## VM Image [deprecated]
+
+In a VM image mode, you run your own operating system (for now only linux is supported)
+The image provided must be
+
+- EFI bootable
+- Cloud-init enabled.
+
+You can find later in this document how to create your own bootable image.
+
+A VM reservations must also have at least 1 volume, as the boot image
+will be copied to this volume. The size of the root disk will be the size of this
+volume.
+
+The image used to the boot the VM must has cloud-init enabled on boot. Cloud-init
+receive its config over the NoCloud source. This takes care of setting up networking, hostname
+, root authorized_keys.
+
+> This method of building a full VM from a raw image is not recommended and will get phased out in
+the future. It's better to use either the container method to run containerized Apps. Another option
+is to run your own kernel from an flist (explained below)
+
+### Expected Flist structure
+
+An `Zmachine` will be considered a `VM` if it contains an `/image.raw` file.
+
+`/image.raw` is used as "boot disk". This `/image.raw` is copied to the first attached
+volume of the `VM`. Cloud-init will take care of resizing the filesystem on the image
+to take the full disk size allocated in the deployment.
+
+Note if the `image.raw` size is larger than the allocated disk. the workload for the VM
+will fail.
+
+### Expected Flist structure
+
+Any Flist will boot as a container, **UNLESS** is has a `/image.raw` file. There is
+no need to specify a kernel yourself (it will be provided).
+
+### Known issues
+
+- We need to do proper performance testing for `virtio-fs`. There seems to be some
+ suboptimal performance right now.
+- It's not currently possible to get container logs.
+- TODO: more testing
+
+## Creating VM image
+
+This is a simple tutorial on how to create your own VM image
+> Note: Please consider checking the official vm images repo on the hub before building your own
+image. this can save you a lot of time (and network traffic) here
+
+### Use one of ubuntu cloud-images
+
+If the ubuntu images in the official repo are not enough, you can simply upload one of the official images as follows
+
+- Visit
+- Select the version you want (let's assume bionic)
+- Go to bionic, then click on current
+- download the amd64.img file like this one
+- This is a `Qcow2` image, this is not supported by zos. So we need to convert this to a raw disk image using the following command
+
+```bash
+qemu-img convert -p -f qcow2 -O raw bionic-server-cloudimg-amd64.img image.raw
+```
+
+- now we have the raw image (image.raw) time to compress and upload to the hub
+
+```bash
+tar -czf ubuntu-18.04-lts.tar.gz image.raw
+```
+
+- now visit the hub and login or create your own account, then click on upload my file button
+- Select the newly created tar.gz file
+- Now you should be able to use this flist to create Zmachine workloads
+
+### Create an image from scratch
+
+This is an advanced scenario and you will require some prior knowledge of how to create local VMs and how to prepare the installation medium,
+and installing your OS of choice.
+
+Before we continue you need to have some hypervisor that you can use locally. Libvirt/Qemu are good choices. Hence we skip on what you need to do to install and configure your system correctly not how to create the VM
+
+#### VM Requirements
+
+Create a VM with enough CPU and Memory to handle the installation process note that this does not relate on what your choices for CPU and Memory are going to be for the actual VM running on the grid.
+
+We going to install arch linux image. So we will have to create a VM with
+
+- Disk of about 2GB (note this also is not related to the final VM running on the grid, on installation the OS image will expand to use the entire allocated disk attached to the VM eventually). The smaller the disk is better this can be different for each OS.
+- Add the arch installation iso or any other installation medium
+
+#### Boot the VM (locally)
+
+Boot the VM to start installation. The boot must support EFI booting because ZOS only support images with esp partition. So make sure that both your hypervisor and boot/installation medium supports this.
+
+For example in Libvirt Manager make sure you are using the right firmware (UEFI)
+
+#### Installation
+
+We going to follow the installation manual for Arch linux but with slight tweaks:
+
+- Make sure VM is booted with UEFI, run `efivar -l` command see if you get any output. Otherwise the machine is probably booted in BIOS mode.
+- With `parted` create 2 partitions
+ - an esp (boot) partition of 100M
+ - a root partition that spans the remaining of the disk
+
+```bash
+DISK=/dev/vda
+# First, create a gpt partition table
+parted $DISK mklabel gpt
+# Secondly, create the esp partition of 100M
+parted $DISK mkpart primary 1 100M
+# Mark first part as esp
+parted $DISK set 1 esp on
+# Use the remaining part as root that takes the remaining
+# space on disk
+parted $DISK mkpart primary 100M 100%
+
+# To verify everything is correct do
+parted $DISK print
+
+# this should 2 partitions the first one is slightly less that 100M and has flags (boot, esp), the second one takes the remaining space
+```
+
+We need to format the partitions as follows:
+
+```bash
+# this one has to be vfat of size 32 as follows
+mkfs.vfat -F 32 /dev/vda1
+# This one can be anything based on your preference as long as it's supported by you OS kernel. we going with ext4 in this tutorial
+mkfs.ext4 -L cloud-root /dev/vda2
+```
+
+Note the label assigned to the /dev/vda2 (root) partition this can be anything but it's needed to configure the boot later when installing the boot loader. Otherwise you can use the partition UUID.
+
+Next, we need to mount the disks
+
+```bash
+mount /dev/vda2 /mnt
+mkdir /mnt/boot
+mount /dev/vda1 /mnt/boot
+```
+
+After disks are mounted as above, we need to start the installation
+
+```bash
+pacstrap /mnt base linux linux-firmware vim openssh cloud-init cloud-guest-utils
+```
+
+This will install basic arch linux but will also include cloud-init, cloud-guest-utils, openssh, and vim for convenience.
+
+Following the installation guid to generate fstab file
+
+```
+genfstab -U /mnt >> /mnt/etc/fstab
+```
+
+And arch-chroot into /mnt `arch-chroot /mnt` to continue the setup. please follow all steps in the installation guide to set timezone, and locales as needed.
+
+- You don't have to set the hostname, this will be setup later on zos when the VM is deployed via cloud-init
+- let's drop the root password all together since login to the VM over ssh will require key authentication only, you can do this by running
+
+```bash
+passwd -d root
+```
+
+We make sure required services are enabled
+
+```bash
+systemctl enable sshd
+systemctl enable systemd-networkd
+systemctl enable systemd-resolved
+systemctl enable cloud-init
+systemctl enable cloud-final
+
+# make sure we using resolved
+rm /etc/resolv.conf
+ln -s /run/systemd/resolve/stub-resolv.conf /etc/resolv.conf
+```
+
+Finally installing the boot loader as follows
+> Only grub2 has been tested and known to work.
+
+```bash
+pacman -S grub
+```
+
+Then we need to install grub
+
+```
+grub-install --target=x86_64-efi --efi-directory=esp --removable
+```
+
+Change default values as follows
+
+```
+vim /etc/default/grub
+```
+
+And make sure to change `GRUB_CMDLINE_LINUX_DEFAULT` as follows
+
+```
+GRUB_CMDLINE_LINUX_DEFAULT="loglevel=3 console=tty console=ttyS0"
+```
+
+> Note: we removed the `quiet` and add the console flags.
+
+Also set the `GRUB_TIMEOUT` to 0 for a faster boot
+
+```
+GRUB_TIMEOUT=0
+```
+
+Then finally generating the config
+
+```
+grub-mkconfig -o /boot/grub/grub.cfg
+```
+
+Last thing we need to do is clean up
+
+- pacman cache by running `rm -rf /var/cache/pacman/pkg`
+- cloud-init state by running `cloud-init clean`
+
+Click `Ctrl+D` to exit the change root, then power off by running `poweroff` command.
+
+> NOTE: if you booted the machine again you always need to do `cloud-init clean` as long as it's not yet deployed on ZOS this to make sure the image has a clean state
+>
+#### Converting the disk
+
+Based on your hypervisor of choice you might need to convert the disk to a `raw` image same way we did with ubuntu image.
+
+```bash
+# this is an optional step in case you used a qcoq disk for the installation. If the disk is already `raw` you can skip this
+qemu-img convert -p -f qcow2 -O raw /path/to/vm/disk.img image.raw
+```
+
+Compress and tar the image.raw as before, and upload to the hub.
+
+```
+tar -czf arch-linux.tar.gz image.raw
+```
diff --git a/collections/documentation/developers/internals/zos/manual/zmount/readme.md b/collections/documentation/developers/internals/zos/manual/zmount/readme.md
new file mode 100644
index 0000000..e7de260
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/manual/zmount/readme.md
@@ -0,0 +1,2 @@
+# `zmount` type
+A `zmount` is a local disk that can be attached directly to a container or a virtual machine. `zmount` only require `size` as input as defined [here](https://github.com/threefoldtech/zos/blob/main/pkg/gridtypes/zos/zmount.go) this workload type is only utilized via the `zmachine` workload.
diff --git a/collections/documentation/developers/internals/zos/performance/cpubench.md b/collections/documentation/developers/internals/zos/performance/cpubench.md
new file mode 100644
index 0000000..2d3f8a7
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/performance/cpubench.md
@@ -0,0 +1,85 @@
+ CPUBenchmark
+
+Table of Contents
+
+- [Overview](#overview)
+- [Configuration](#configuration)
+- [Details](#details)
+- [Result Sample](#result-sample)
+- [Result Explanation](#result-explanation)
+
+***
+
+## Overview
+
+The `CPUBenchmark` task is designed to measure the performance of the CPU. it utilizes the [cpu-benchmark-simple](https://github.com/threefoldtech/cpu-benchmark-simple) tool and includes a zos stub to gather the number of workloads running on the node.
+
+## Configuration
+
+- Name: `cpu-benchmark`
+- Schedule: 4 times a day
+- Jitter: 0
+
+## Details
+
+- The benchmark simply runs a `CRC64` computation task, calculates the time spent in the computation and reports it in `seconds`.
+- The computation is performed in both single-threaded and multi-threaded scenarios.
+- Lower time = better performance: for a single threaded benchmark, a lower execution time indicates better performance.
+
+## Result Sample
+
+```json
+{
+ "description": "Measures the performance of the node CPU by reporting the time spent of computing a task in seconds.",
+ "name": "cpu-benchmark",
+ "result": {
+ "multi": 1.105,
+ "single": 1.135,
+ "threads": 1,
+ "workloads": 0
+ },
+ "timestamp": 1700504403
+}
+```
+
+## Result Explanation
+
+The best way to know what's a good or bad value is by testing and comparing different hardware.
+Here are some examples:
+
+**1x Intel(R) Xeon(R) W-2145 CPU @ 3.70GHz** (Q3'2017)
+
+```
+Single thread score: 0.777
+Multi threads score: 13.345 [16 threads]
+```
+
+**1x Intel(R) Pentium(R) CPU G4400 @ 3.30GHz** (Q3'2015)
+
+```
+Single thread score: 1.028
+Multi threads score: 2.089 [2 threads]
+```
+
+**1x Intel(R) Core(TM) i5-3570 CPU @ 3.40GHz** (Q2'2012)
+
+```
+Single thread score: 2.943
+Multi threads score: 12.956 [4 threads]
+```
+
+**2x Intel(R) Xeon(R) CPU E5-2630 v3 @ 2.40GHz** (Q1'2012)
+
+```
+Single thread score: 1.298
+Multi threads score: 44.090 [32 threads]
+```
+
+**2x Intel(R) Xeon(R) CPU L5640 @ 2.27GHz** (Q1'2010)
+
+```
+Single thread score: 2.504
+Multi threads score: 72.452 [24 threads]
+```
+
+As you can see, the more recent the CPU is, the faster it is, but for a same launch period, you can see Xeon way better than regular/desktop CPU. You have to take in account the amount of threads and the time per threads.
\ No newline at end of file
diff --git a/collections/documentation/developers/internals/zos/performance/healthcheck.md b/collections/documentation/developers/internals/zos/performance/healthcheck.md
new file mode 100644
index 0000000..a41c059
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/performance/healthcheck.md
@@ -0,0 +1,38 @@
+ Health Check
+
+Table of Contents
+
+- [Overview](#overview)
+- [Configuration](#configuration)
+- [Details](#details)
+- [Result Sample](#result-sample)
+
+***
+
+## Overview
+
+Health check task executes some checks over ZOS components to determine if the node is in a usable state or not and set flags for the Power Daemon to stop uptime reports if the node is unusable.
+
+## Configuration
+
+- Name: `healthcheck`
+- Schedule: Every 20 mins.
+
+## Details
+
+- Check if the node cache disk is usable or not by trying to write some data to it. If it failed, it set the Readonly flag.
+
+## Result Sample
+
+```json
+{
+ "description": "health check task runs multiple checks to ensure the node is in a usable state and set flags for the power daemon to stop reporting uptime if it is not usable",
+ "name": "healthcheck",
+ "result": {
+ "cache": [
+ "failed to write to cache: open /var/cache/healthcheck: operation not permitted"
+ ]
+ },
+ "timestamp": 1701599580
+}
+```
\ No newline at end of file
diff --git a/collections/documentation/developers/internals/zos/performance/iperf.md b/collections/documentation/developers/internals/zos/performance/iperf.md
new file mode 100644
index 0000000..d7a36dc
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/performance/iperf.md
@@ -0,0 +1,80 @@
+ IPerf
+
+Table of Contents
+
+- [Overview](#overview)
+- [Configuration](#configuration)
+- [Details](#details)
+- [Result Sample](#result-sample)
+
+***
+
+## Overview
+
+The `iperf` package is designed to facilitate network performance testing using the `iperf3` tool. with both UDP and TCP over IPv4 and IPv6.
+
+## Configuration
+
+- Name: `iperf`
+- Schedule: 4 times a day
+- Jitter: 20 min
+
+## Details
+
+- The package using the iperf binary to examine network performance under different conditions.
+- It randomly fetch PublicConfig data for randomly public nodes on the chain + all public node from free farm. These nodes serve as the targets for the iperf tests.
+- For each node, it run the test with 4 times. through (UDP/TCP) using both node IPs (v4/v6)
+- result will be a slice of all public node report (4 for each) each one will include:
+ ```
+ UploadSpeed: Upload speed (in bits per second).
+ DownloadSpeed: Download speed (in bits per second).
+ NodeID: ID of the node where the test was conducted.
+ NodeIpv4: IPv4 address of the node.
+ TestType: Type of the test (TCP or UDP).
+ Error: Any error encountered during the test.
+ CpuReport: CPU utilization report (in percentage).
+ ```
+
+## Result Sample
+
+```json
+{
+ "description": "Test public nodes network performance with both UDP and TCP over IPv4 and IPv6",
+ "name": "iperf",
+ "result": [
+ {
+ "cpu_report": {
+ "host_system": 2.4433388913571044,
+ "host_total": 3.542919199613454,
+ "host_user": 1.0996094859359695,
+ "remote_system": 0.24430594945859846,
+ "remote_total": 0.3854457128784448,
+ "remote_user": 0.14115962407747246
+ },
+ "download_speed": 1041274.4792242317,
+ "error": "",
+ "node_id": 124,
+ "node_ip": "88.99.30.200",
+ "test_type": "tcp",
+ "upload_speed": 1048549.3668460822
+ },
+ {
+ "cpu_report": {
+ "host_system": 0,
+ "host_total": 0,
+ "host_user": 0,
+ "remote_system": 0,
+ "remote_total": 0,
+ "remote_user": 0
+ },
+ "download_speed": 0,
+ "error": "unable to connect to server - server may have stopped running or use a different port, firewall issue, etc.: Network unreachable",
+ "node_id": 124,
+ "node_ip": "2a01:4f8:10a:710::2",
+ "test_type": "tcp",
+ "upload_speed": 0
+ }
+ ],
+ "timestamp": 1700507035
+}
+```
\ No newline at end of file
diff --git a/collections/documentation/developers/internals/zos/performance/performance.md b/collections/documentation/developers/internals/zos/performance/performance.md
new file mode 100644
index 0000000..7f3ea76
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/performance/performance.md
@@ -0,0 +1,90 @@
+ Performance Monitor Package
+
+Table of Contents
+
+- [Overview](#overview)
+- [Flow](#flow)
+- [Node Initialization Check](#node-initialization-check)
+- [Scheduling](#scheduling)
+- [RMB Commands](#rmb-commands)
+- [Caching](#caching)
+- [Registered Tests](#registered-tests)
+- [Test Suite](#test-suite)
+
+***
+
+## Overview
+
+The `perf` package is a performance monitor in `zos` nodes. it schedules tasks, cache their results and allows retrieval of these results through `RMB` calls.
+
+## Flow
+
+1. The `perf` monitor is started by the `noded` service in zos.
+2. Tasks are registered with a schedule in the new monitor.
+3. A bus handler is opened to allow result retrieval.
+
+## Node Initialization Check
+
+To ensure that the node always has a test result available, a check is performed on node startup for all the registered tasks, if a task doesn't have any stored result, it will run immediately without waiting for the next scheduled time.
+
+## Scheduling
+
+- Tasks are scheduled using a 6 fields cron format. this format provides flexibility to define time, allowing running tasks periodically or at specific time.
+
+- Each task has a jitter which is the maximum number of seconds the task could sleep before it runs, this happens to prevent all tests ending up running at exactly the same time. So, for example, if a task is scheduled to run at `06:00` and its jitter is `10`, it is expected to run anywhere between `06:00` and `06:10`.
+
+## RMB Commands
+
+- `zos.perf.get`:
+
+ - Payload: a payload type that contains the name of the test
+
+ ```go
+ type Payload struct {
+ Name string
+ }
+ ```
+
+ Possible values:
+
+ - `"public-ip-validation"`
+ - `"cpu-benchmark"`
+ - `"iperf"`
+
+ - Return: a single task result.
+
+ - Possible Error: `ErrResultNotFound` if no result is stored for the given task.
+
+- `zos.perf.get_all`:
+
+ - Return: all stored results
+
+The rmb direct client can be used to call these commands. check the [example](https://github.com/threefoldtech/tfgrid-sdk-go/blob/development/rmb-sdk-go/examples/rpc_client/main.go)
+
+## Caching
+
+Results are stored in a Redis server running on the node.
+
+The key in redis is the name of the task prefixed with the word `perf`.
+The value is an instance of `TaskResult` struct contains:
+
+- Name of the task
+- Timestamp when the task was run
+- A brief description about what the task do
+- The actual returned result from the task
+
+Notes:
+
+- Storing results by a key ensures each new result overrides the old one, so there is always a single result for each task.
+- Storing results prefixed with `perf` eases retrieving all the results stored by this module.
+
+## Registered Tests
+
+- [Public IP Validation](./publicips.md)
+- [CPUBenchmark](./cpubench.md)
+- [IPerf](./iperf.md)
+- [Health Check](./healthcheck.md)
+
+## Test Suite
+
+Go to [this link](https://app.testlodge.com/a/26076/projects/40893/suites/234919) for a test suite covering the test cases for the performance testing.
\ No newline at end of file
diff --git a/collections/documentation/developers/internals/zos/performance/publicips.md b/collections/documentation/developers/internals/zos/performance/publicips.md
new file mode 100644
index 0000000..0549512
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/performance/publicips.md
@@ -0,0 +1,55 @@
+ Public IPs Validation Task
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Configuration](#configuration)
+- [Task Details](#task-details)
+- [Result](#result)
+ - [Result Sample](#result-sample)
+
+***
+
+## Introduction
+
+The goal of the task is to make sure public IPs assigned to a farm are valid and can be assigned to deployments.
+
+## Configuration
+
+- Name: `public-ip-validation`
+- Schedule: 4 times a day
+- Jitter: 10 min
+
+## Task Details
+
+- The task depends on `Networkd` ensuring the proper test network setup is correct and will fail if it wasn't setup properly. The network setup consists of a test Namespace and a MacVLAN as part of it. All steps are done inside the test Namespace.
+- Decide if the node should run the task or another one in the farm based on the node ID. The node with the least ID and with power target as up should run it. The other will log why they shouldn't run the task and return with no errors. This is done to ensure only one node runs the task to avoid problems like assigning the same IP.
+- Get public IPs set on the farm.
+- Remove all IPs and routes added to the test MacVLAN to ensure any remaining from previous task run are removed.
+- Skip IPs that are assigned to a contract.
+- Set the MacVLAN link up.
+- Iterate over all public IPs and add them with the provided gateway to the MacVLAN.
+- Validate the IP by querying an external source that return the public IP for the node.
+- If the public IP returned matches the IP added in the link, then the IP is valid. Otherwise, it is invalid.
+- Remove all IPs and routes between each IP to make them available for other deployments.
+- After iterating over all public IPs, set the link down.
+
+## Result
+
+The task only returns a single map of String (IP) to IPReport. The report consists of the IP state (valid, invalid or skipped) and the reason for the state.
+
+### Result Sample
+
+```json
+{
+ "description": "Runs on the least NodeID node in a farm to validate all its IPs.",
+ "name": "public-ip-validation",
+ "result": {
+ "185.206.122.29/24": {
+ "reason": "public ip or gateway data are not valid",
+ "state": "invalid"
+ }
+ },
+ "timestamp": 1700504421
+}
+```
\ No newline at end of file
diff --git a/collections/documentation/developers/internals/zos/readme.md b/collections/documentation/developers/internals/zos/readme.md
new file mode 100644
index 0000000..33d16df
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/readme.md
@@ -0,0 +1,28 @@
+ Zero-OS
+
+ Table of Contents
+
+- [Manual](./manual/manual.md)
+- [Workload Types](./manual/workload_types.md)
+- [Internal Modules](./internals/internals.md)
+ - [Identity](./internals/identity/index.md)
+ - [Node ID Generation](./internals/identity/identity.md)
+ - [Node Upgrade](./internals/identity/upgrade.md)
+ - [Node](./internals/node/index.md)
+ - [Storage](./internals/storage/index.md)
+ - [Network](./internals/network/index.md)
+ - [Introduction](./internals/network/introduction.md)
+ - [Definitions](./internals/network/definitions.md)
+ - [Mesh](./internals/network/mesh.md)
+ - [Setup](./internals/network/setup_farm_network.md)
+ - [Flist](./internals/flist/index.md)
+ - [Container](./internals/container/index.md)
+ - [VM](./internals/vmd/index.md)
+ - [Provision](./internals/provision/index.md)
+- [Capacity](./internals/capacity.md)
+- [Performance Monitor Package](./performance/performance.md)
+ - [Public IPs Validation Task](./performance/publicips.md)
+ - [CPUBenchmark](./performance/cpubench.md)
+ - [IPerf](./performance/iperf.md)
+ - [Health Check](./performance/healthcheck.md)
+- [API](./manual/api.md)
diff --git a/collections/documentation/developers/internals/zos/release/readme.md b/collections/documentation/developers/internals/zos/release/readme.md
new file mode 100644
index 0000000..6af1a51
--- /dev/null
+++ b/collections/documentation/developers/internals/zos/release/readme.md
@@ -0,0 +1,31 @@
+# Releases of Zero-OS
+
+We use a simple pipeline release workflow. Building and file distribution are made using GitHub Actions.
+Usable files are available on the [Zero-OS Hub](https://hub.grid.tf/tf-zos).
+
+This pipeline is made to match the 3 different type of running mode of 0-OS. For more information head to the [upgrade documentation](../identity/upgrade.md).
+
+## Development build
+
+On a push to main branch on the zos repository, a new development build is triggered. If the build succeed,
+binaries are packed into an flist and uploaded to the [tf-autobuilder](https://hub.grid.tf/tf-autobuilder) repository of the hub.
+
+This flist is then promoted into the [tf-zos](https://hub.grid.tf/tf-zos) repository of the hub and a symlink to this latest build is made (`tf-autobuilder/zos:development-3:latest.flist`)
+
+## Releases
+We create 3 types of releases:
+- QA release, in this release the version is suffixed by `qa` for example `v3.5.0-qa1`.
+- RC release, in this release the version is suffixed by `rc` for example `v3.5.0-rc2`.
+- Main release, is this release the version has no suffix, for example `v3.5.0`
+
+The release cycle goes like this:
+- As mentioned before devnet is updated the moment new code is available on `main` branch. Since the `dev` release is auto linked to the latest `flist` on the hub. Nodes on devnet will auto update to the latest available build.
+- Creating a `qa` release, will not not trigger the same behavior on `qa` net, same for both testnet and mainnet. Instead a workflow must be triggered, this is only to make sure 100% that an update is needed.
+- Once the build of the release is available, a [deploy](../../.github/workflows/grid-deploy.yaml) workflow needed to be triggered with the right version to deploy on the proper network.
+ - The work flow all what it does is linking the right version under the hub [tf-zos](https://hub.grid.tf/tf-zos) repo
+
+> The `deploy` flow is rarely used, the on chain update is also available. By setting the right version on tfchain, the link on the hub is auto-updated and hence the deploy workflow won't be needed to be triggered. Although we have it now as a safety net in case something goes wrong (chain is broken) and we need to force a specific version on ZOS.
+
+- Development: https://playground.hub.grid.tf/tf-autobuilder/zos:development-3:latest.flist
+- Testing: https://playground.hub.grid.tf/tf-zos/zos:testing-3:latest.flist
+- Production: https://playground.hub.grid.tf/tf-zos/zos:production-3:latest.flist
diff --git a/collections/documentation/developers/javascript/grid3_javascript_capacity_planning.md b/collections/documentation/developers/javascript/grid3_javascript_capacity_planning.md
new file mode 100644
index 0000000..f1db3cc
--- /dev/null
+++ b/collections/documentation/developers/javascript/grid3_javascript_capacity_planning.md
@@ -0,0 +1,110 @@
+ Capacity Planning
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Example](#example)
+
+***
+
+## Introduction
+
+It's almost the same as in [deploying a single VM](../javascript/grid3_javascript_vm.md) the only difference is you can automate the choice of the node to deploy on using code. We now support `FilterOptions` to filter nodes based on specific criteria e.g the node resources (CRU, SRU, HRU, MRU) or being part of a specific farm or located in some country, or being a gateway or not
+
+## Example
+
+```ts
+FilterOptions: { accessNodeV4?: boolean; accessNodeV6?: boolean; city?: string; country?: string; cru?: number; hru?: number; mru?: number; sru?: number; farmId?: number; farmName?: string; gateway?: boolean; publicIPs?: boolean; certified?: boolean; dedicated?: boolean; availableFor?: number; page?: number;}
+```
+
+```ts
+import { DiskModel, FilterOptions, MachineModel, MachinesModel, NetworkModel } from "../src";
+import { config, getClient } from "./client_loader";
+import { log } from "./utils";
+
+async function main() {
+ const grid3 = await getClient();
+
+ // create network Object
+ const n = new NetworkModel();
+ n.name = "dynamictest";
+ n.ip_range = "10.249.0.0/16";
+
+ // create disk Object
+ const disk = new DiskModel();
+ disk.name = "dynamicDisk";
+ disk.size = 8;
+ disk.mountpoint = "/testdisk";
+
+ const vmQueryOptions: FilterOptions = {
+ cru: 1,
+ mru: 2, // GB
+ sru: 9,
+ country: "Belgium",
+ availableFor: grid3.twinId,
+ };
+
+ // create vm node Object
+ const vm = new MachineModel();
+ vm.name = "testvm";
+ vm.node_id = +(await grid3.capacity.filterNodes(vmQueryOptions))[0].nodeId; // TODO: allow random choise
+ vm.disks = [disk];
+ vm.public_ip = false;
+ vm.planetary = true;
+ vm.cpu = 1;
+ vm.memory = 1024 * 2;
+ vm.rootfs_size = 0;
+ vm.flist = "https://hub.grid.tf/tf-official-apps/base:latest.flist";
+ vm.entrypoint = "/sbin/zinit init";
+ vm.env = {
+ SSH_KEY: config.ssh_key,
+ };
+
+ // create VMs Object
+ const vms = new MachinesModel();
+ vms.name = "dynamicVMS";
+ vms.network = n;
+ vms.machines = [vm];
+ vms.metadata = "{'testVMs': true}";
+ vms.description = "test deploying VMs via ts grid3 client";
+
+ // deploy vms
+ const res = await grid3.machines.deploy(vms);
+ log(res);
+
+ // get the deployment
+ const l = await grid3.machines.getObj(vms.name);
+ log(l);
+
+ // // delete
+ // const d = await grid3.machines.delete({ name: vms.name });
+ // log(d);
+
+ await grid3.disconnect();
+}
+
+main();
+```
+
+In this example you can notice the criteria for `server1`
+
+```typescript
+const server1_options: FilterOptions = {
+ cru: 1,
+ mru: 2, // GB
+ sru: 9,
+ country: "Belgium",
+ availableFor: grid3.twinId,
+};
+
+```
+
+Here we want all the nodes with `CRU:1`, `MRU:2`, `SRU:9`, located in `Belgium` and available for me (not rented for someone else).
+
+> Note some libraries allow reverse lookup of countries codes by name e.g [i18n-iso-countries](https://www.npmjs.com/package/i18n-iso-countries)
+
+and then in the MachineModel, we specified the `node_id` to be the first value of our filteration
+
+```typescript
+vm.node_id = +(await nodes.filterNodes(server1_options))[0].nodeId;
+```
diff --git a/collections/documentation/developers/javascript/grid3_javascript_caprover.md b/collections/documentation/developers/javascript/grid3_javascript_caprover.md
new file mode 100644
index 0000000..1b1e1e3
--- /dev/null
+++ b/collections/documentation/developers/javascript/grid3_javascript_caprover.md
@@ -0,0 +1,232 @@
+ Deploy CapRover
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Leader Node](#leader-node)
+ - [Code Example](#code-example)
+ - [Environment Variables](#environment-variables)
+- [Worker Node](#worker-node)
+ - [Code Example](#code-example-1)
+ - [Environment Variables](#environment-variables-1)
+- [Questions and Feedback](#questions-and-feedback)
+
+***
+
+## Introduction
+
+In this section, we show how to deploy CapRover with the Javascript client.
+
+This deployment is very similar to what we have in the section [Deploy a VM](./grid3_javascript_vm.md), but the environment variables are different.
+
+## Leader Node
+
+We present here a code example and the environment variables to deploy a CapRover Leader node.
+
+For further details about the Leader node deployment, [read this documentation](https://github.com/freeflowuniverse/freeflow_caprover#a-leader-node-deploymentsetup).
+
+### Code Example
+
+```ts
+import {
+ DiskModel,
+ FilterOptions,
+ MachineModel,
+ MachinesModel,
+ NetworkModel,
+} from "../src";
+import { config, getClient } from "./client_loader";
+import { log } from "./utils";
+
+async function main() {
+ const grid3 = await getClient();
+
+ const vmQueryOptions: FilterOptions = {
+ cru: 4,
+ mru: 4, // GB
+ sru: 10,
+ farmId: 1,
+ };
+
+ const CAPROVER_FLIST =
+ "https://hub.grid.tf/tf-official-apps/tf-caprover-latest.flist";
+ // create network Object
+ const n = new NetworkModel();
+ n.name = "wedtest";
+ n.ip_range = "10.249.0.0/16";
+
+ // create disk Object
+ const disk = new DiskModel();
+ disk.name = "wedDisk";
+ disk.size = 10;
+ disk.mountpoint = "/var/lib/docker";
+
+ // create vm node Object
+ const vm = new MachineModel();
+ vm.name = "testvm";
+ vm.node_id = +(await grid3.capacity.filterNodes(vmQueryOptions))[0].nodeId;
+ vm.disks = [disk];
+ vm.public_ip = true;
+ vm.planetary = false;
+ vm.cpu = 4;
+ vm.memory = 1024 * 4;
+ vm.rootfs_size = 0;
+ vm.flist = CAPROVER_FLIST;
+ vm.entrypoint = "/sbin/zinit init";
+ vm.env = {
+ PUBLIC_KEY: config.ssh_key,
+ SWM_NODE_MODE: "leader",
+ CAPROVER_ROOT_DOMAIN: "rafy.grid.tf", // update me
+ DEFAULT_PASSWORD: "captain42",
+ CAPTAIN_IMAGE_VERSION: "latest",
+ };
+
+ // create VMs Object
+ const vms = new MachinesModel();
+ vms.name = "newVMS5";
+ vms.network = n;
+ vms.machines = [vm];
+ vms.metadata = "{'testVMs': true}";
+ vms.description = "caprover leader machine/node";
+
+ // deploy vms
+ const res = await grid3.machines.deploy(vms);
+ log(res);
+
+ // get the deployment
+ const l = await grid3.machines.getObj(vms.name);
+ log(l);
+
+ log(
+ `You can access Caprover via the browser using: https://captain.${vm.env.CAPROVER_ROOT_DOMAIN}`
+ );
+
+ // // delete
+ // const d = await grid3.machines.delete({ name: vms.name });
+ // log(d);
+
+ await grid3.disconnect();
+}
+
+main();
+```
+
+
+
+### Environment Variables
+
+- PUBLIC_KEY: Your public IP to be able to access the VM.
+- SWM_NODE_MODE: Caprover Node type which must be `leader` as we are deploying a leader node.
+- CAPROVER_ROOT_DOMAIN: The domain which you we will use to bind the deployed VM.
+- DEFAULT_PASSWORD: Caprover default password you want to deploy with.
+
+
+
+## Worker Node
+
+We present here a code example and the environment variables to deploy a CapRover Worker node.
+
+Note that before deploying the Worker node, you should check the following:
+
+- Get the Leader node public IP address.
+- The Worker node should join the cluster from the UI by adding public IP address and the private SSH key.
+
+For further information, [read this documentation](https://github.com/freeflowuniverse/freeflow_caprover#step-4-access-the-captain-dashboard).
+
+### Code Example
+
+```ts
+import {
+ DiskModel,
+ FilterOptions,
+ MachineModel,
+ MachinesModel,
+ NetworkModel,
+} from "../src";
+import { config, getClient } from "./client_loader";
+import { log } from "./utils";
+
+async function main() {
+ const grid3 = await getClient();
+
+ const vmQueryOptions: FilterOptions = {
+ cru: 4,
+ mru: 4, // GB
+ sru: 10,
+ farmId: 1,
+ };
+
+ const CAPROVER_FLIST =
+ "https://hub.grid.tf/tf-official-apps/tf-caprover-latest.flist";
+ // create network Object
+ const n = new NetworkModel();
+ n.name = "wedtest";
+ n.ip_range = "10.249.0.0/16";
+
+ // create disk Object
+ const disk = new DiskModel();
+ disk.name = "wedDisk";
+ disk.size = 10;
+ disk.mountpoint = "/var/lib/docker";
+
+ // create vm node Object
+ const vm = new MachineModel();
+ vm.name = "capworker1";
+ vm.node_id = +(await grid3.capacity.filterNodes(vmQueryOptions))[0].nodeId;
+ vm.disks = [disk];
+ vm.public_ip = true;
+ vm.planetary = false;
+ vm.cpu = 4;
+ vm.memory = 1024 * 4;
+ vm.rootfs_size = 0;
+ vm.flist = CAPROVER_FLIST;
+ vm.entrypoint = "/sbin/zinit init";
+ vm.env = {
+ // These env. vars needed to be changed based on the leader node.
+ PUBLIC_KEY: config.ssh_key,
+ SWM_NODE_MODE: "worker",
+ LEADER_PUBLIC_IP: "185.206.122.157",
+ CAPTAIN_IMAGE_VERSION: "latest",
+ };
+
+ // create VMs Object
+ const vms = new MachinesModel();
+ vms.name = "newVMS6";
+ vms.network = n;
+ vms.machines = [vm];
+ vms.metadata = "{'testVMs': true}";
+ vms.description = "caprover worker machine/node";
+
+ // deploy vms
+ const res = await grid3.machines.deploy(vms);
+ log(res);
+
+ // get the deployment
+ const l = await grid3.machines.getObj(vms.name);
+ log(l);
+
+ // // delete
+ // const d = await grid3.machines.delete({ name: vms.name });
+ // log(d);
+
+ await grid3.disconnect();
+}
+
+main();
+```
+
+
+
+### Environment Variables
+
+The deployment of the Worker node is similar to the deployment of the Leader node, with the exception of the environment variables which differ slightly.
+
+- PUBLIC_KEY: Your public IP to be able to access the VM.
+- SWM_NODE_MODE: Caprover Node type which must be `worker` as we are deploying a worker node.
+- LEADER_PUBLIC_IP: Leader node public IP.
+
+
+
+## Questions and Feedback
+
+If you have any questions, you can ask the ThreeFold community for help on the [ThreeFold Forum](http://forum.threefold.io/) or on the [ThreeFold Grid Tester Community](https://t.me/threefoldtesting) on Telegram.
\ No newline at end of file
diff --git a/collections/documentation/developers/javascript/grid3_javascript_gpu_support.md b/collections/documentation/developers/javascript/grid3_javascript_gpu_support.md
new file mode 100644
index 0000000..4c56ca6
--- /dev/null
+++ b/collections/documentation/developers/javascript/grid3_javascript_gpu_support.md
@@ -0,0 +1,91 @@
+ GPU Support and JavaScript
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Example](#example)
+
+***
+
+## Introduction
+
+We present here a quick introduction to GPU support with JavaScript.
+
+There are a couple of updates regarding finding nodes with GPU, querying node for GPU information and deploying with support of GPU.
+
+This is an ongoing development and this section will be updated as new information comes in.
+
+## Example
+
+Here is an example script to deploy with GPU support:
+
+```ts
+import { DiskModel, FilterOptions, MachineModel, MachinesModel, NetworkModel } from "../src";
+import { config, getClient } from "./client_loader";
+import { log } from "./utils";
+
+async function main() {
+ const grid3 = await getClient();
+
+ // create network Object
+ const n = new NetworkModel();
+ n.name = "vmgpuNetwork";
+ n.ip_range = "10.249.0.0/16";
+
+ // create disk Object
+ const disk = new DiskModel();
+ disk.name = "vmgpuDisk";
+ disk.size = 100;
+ disk.mountpoint = "/testdisk";
+
+ const vmQueryOptions: FilterOptions = {
+ cru: 8,
+ mru: 16, // GB
+ sru: 100,
+ availableFor: grid3.twinId,
+ hasGPU: true,
+ rentedBy: grid3.twinId,
+ };
+
+ // create vm node Object
+ const vm = new MachineModel();
+ vm.name = "vmgpu";
+ vm.node_id = +(await grid3.capacity.filterNodes(vmQueryOptions))[0].nodeId; // TODO: allow random choice
+ vm.disks = [disk];
+ vm.public_ip = false;
+ vm.planetary = true;
+ vm.cpu = 8;
+ vm.memory = 1024 * 16;
+ vm.rootfs_size = 0;
+ vm.flist = "https://hub.grid.tf/tf-official-vms/ubuntu-22.04.flist";
+ vm.entrypoint = "/";
+ vm.env = {
+ SSH_KEY: config.ssh_key,
+ };
+ vm.gpu = ["0000:0e:00.0/1002/744c"]; // gpu card's id, you can check the available gpu from the dashboard
+
+ // create VMs Object
+ const vms = new MachinesModel();
+ vms.name = "vmgpu";
+ vms.network = n;
+ vms.machines = [vm];
+ vms.metadata = "";
+ vms.description = "test deploying VM with GPU via ts grid3 client";
+
+ // deploy vms
+ const res = await grid3.machines.deploy(vms);
+ log(res);
+
+ // get the deployment
+ const l = await grid3.machines.getObj(vms.name);
+ log(l);
+
+ // delete
+ const d = await grid3.machines.delete({ name: vms.name });
+ log(d);
+
+ await grid3.disconnect();
+}
+
+main();
+```
\ No newline at end of file
diff --git a/collections/documentation/developers/javascript/grid3_javascript_installation.md b/collections/documentation/developers/javascript/grid3_javascript_installation.md
new file mode 100644
index 0000000..3040880
--- /dev/null
+++ b/collections/documentation/developers/javascript/grid3_javascript_installation.md
@@ -0,0 +1,124 @@
+Installation
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Prerequisites](#prerequisites)
+- [Installation](#installation)
+ - [External Package](#external-package)
+ - [Local Usage](#local-usage)
+- [Getting Started](#getting-started)
+ - [Client Configuration](#client-configuration)
+- [Generate the Documentation](#generate-the-documentation)
+- [How to Run the Scripts](#how-to-run-the-scripts)
+- [Reference API](#reference-api)
+
+***
+
+## Introduction
+
+We present here the general steps required to install and use the ThreeFold Grid Client.
+
+The [Grid Client](https://github.com/threefoldtech/tfgrid-sdk-ts/tree/development/packages/grid_client) is written using [TypeScript](https://www.typescriptlang.org/) to provide more convenience and type-checked code. It is used to deploy workloads like virtual machines, kubernetes clusters, quantum storage, and more.
+
+## Prerequisites
+
+To install the Grid Client, you will need the following on your machine:
+
+- [Node.js](https://nodejs.org/en) ^18
+- npm 8.2.0 or higher
+- may need to install libtool (**apt-get install libtool**)
+
+> Note: [nvm](https://nvm.sh/) is the recommended way for installing node.
+
+To use the Grid Client, you will need the following on the TFGrid:
+
+- A TFChain account
+- TFT in your wallet
+
+If it is not the case, please visit the [Get started section](../../system_administrators/getstarted/tfgrid3_getstarted.md).
+
+## Installation
+
+### External Package
+
+To install the external package, simply run the following command:
+
+```bash
+yarn add @threefold/grid_client
+```
+
+> Note: For the **qa**, **test** and **main** networks, please use @2.1.1 version.
+
+### Local Usage
+
+To use the Grid Client locally, clone the repository then install the Grid Client:
+
+- Clone the repository
+ - ```bash
+ git clone https://github.com/threefoldtech/tfgrid-sdk-ts
+ ```
+- Install the Grid Client
+ - With yarn
+ - ```bash
+ yarn install
+ ```
+ - With npm
+ - ```bash
+ npm install
+ ```
+
+> Note: In the directory **grid_client/scripts**, we provided a set of scripts to test the Grid Client.
+
+## Getting Started
+
+You will need to set the client configuration either by setting the json file manually (**scripts/config.json**) or by using the provided script (**scripts/client_loader.ts**).
+
+### Client Configuration
+
+Make sure to set the client configuration properly before using the Grid Client.
+
+- **network**: The network environment (**dev**, **qa**, **test** or **main**).
+
+- **mnemonic**: The 12 words mnemonics for your account.
+ - Learn how to create one [here](../../dashboard/wallet_connector.md).
+
+- **storeSecret**: This is any word that will be used for encrypting/decrypting the keys on ThreeFold key-value store.
+
+- **ssh_key**: The public SSH key set on your machine.
+
+> Note: Only networks can't be isolated, all projects can see the same network.
+
+## Generate the Documentation
+
+The easiest way to test the installation is to run the following command with either yarn or npm to generate the Grid Client documentation:
+
+* With yarn
+ * ```
+ yarn run serve-docs
+ ```
+* With npm
+ * ```
+ npm run serve-docs
+ ```
+
+> Note: You can also use the command **yarn run** to see all available options.
+
+## How to Run the Scripts
+
+You can explore the Grid Client by testing the different scripts proposed in **grid_client/scripts**.
+
+- Update your customized deployments specs if needed
+- Run using [ts-node](https://www.npmjs.com/ts-node)
+ - With yarn
+ - ```bash
+ yarn run ts-node --project tsconfig-node.json scripts/zdb.ts
+ ```
+ - With npx
+ - ```bash
+ npx ts-node --project tsconfig-node.json scripts/zdb.ts
+ ```
+
+## Reference API
+
+While this is still a work in progress, you can have a look [here](https://threefoldtech.github.io/tfgrid-sdk-ts/packages/grid_client/docs/api/index.html).
diff --git a/collections/documentation/developers/javascript/grid3_javascript_kubernetes.md b/collections/documentation/developers/javascript/grid3_javascript_kubernetes.md
new file mode 100644
index 0000000..645f3e2
--- /dev/null
+++ b/collections/documentation/developers/javascript/grid3_javascript_kubernetes.md
@@ -0,0 +1,186 @@
+ Deploying a Kubernetes Cluster
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Prerequisites](#prerequisites)
+- [Example code](#example-code)
+- [Detailed explanation](#detailed-explanation)
+ - [Building network](#building-network)
+ - [Building nodes](#building-nodes)
+ - [Building cluster](#building-cluster)
+ - [Deploying](#deploying)
+ - [Getting deployment information](#getting-deployment-information)
+ - [Deleting deployment](#deleting-deployment)
+
+***
+
+## Introduction
+
+We show how to deploy a Kubernetes cluster on the TFGrid with the Javascript client.
+
+## Prerequisites
+
+- Make sure you have your [client](./grid3_javascript_loadclient.md) prepared
+
+## Example code
+
+```ts
+import { FilterOptions, K8SModel, KubernetesNodeModel, NetworkModel } from "../src";
+import { config, getClient } from "./client_loader";
+import { log } from "./utils";
+
+async function main() {
+ const grid3 = await getClient();
+
+ // create network Object
+ const n = new NetworkModel();
+ n.name = "monNetwork";
+ n.ip_range = "10.238.0.0/16";
+ n.addAccess = true;
+
+ const masterQueryOptions: FilterOptions = {
+ cru: 2,
+ mru: 2, // GB
+ sru: 2,
+ availableFor: grid3.twinId,
+ farmId: 1,
+ };
+
+ const workerQueryOptions: FilterOptions = {
+ cru: 1,
+ mru: 1, // GB
+ sru: 1,
+ availableFor: grid3.twinId,
+ farmId: 1,
+ };
+
+ // create k8s node Object
+ const master = new KubernetesNodeModel();
+ master.name = "master";
+ master.node_id = +(await grid3.capacity.filterNodes(masterQueryOptions))[0].nodeId;
+ master.cpu = 1;
+ master.memory = 1024;
+ master.rootfs_size = 0;
+ master.disk_size = 1;
+ master.public_ip = false;
+ master.planetary = true;
+
+ // create k8s node Object
+ const worker = new KubernetesNodeModel();
+ worker.name = "worker";
+ worker.node_id = +(await grid3.capacity.filterNodes(workerQueryOptions))[0].nodeId;
+ worker.cpu = 1;
+ worker.memory = 1024;
+ worker.rootfs_size = 0;
+ worker.disk_size = 1;
+ worker.public_ip = false;
+ worker.planetary = true;
+
+ // create k8s Object
+ const k = new K8SModel();
+ k.name = "testk8s";
+ k.secret = "secret";
+ k.network = n;
+ k.masters = [master];
+ k.workers = [worker];
+ k.metadata = "{'testk8s': true}";
+ k.description = "test deploying k8s via ts grid3 client";
+ k.ssh_key = config.ssh_key;
+
+ // deploy
+ const res = await grid3.k8s.deploy(k);
+ log(res);
+
+ // get the deployment
+ const l = await grid3.k8s.getObj(k.name);
+ log(l);
+
+ // // delete
+ // const d = await grid3.k8s.delete({ name: k.name });
+ // log(d);
+
+ await grid3.disconnect();
+}
+
+main();
+```
+
+## Detailed explanation
+
+### Building network
+
+```typescript
+// create network Object
+const n = new NetworkModel();
+n.name = "monNetwork";
+n.ip_range = "10.238.0.0/16";
+
+```
+
+### Building nodes
+
+```typescript
+// create k8s node Object
+const master = new KubernetesNodeModel();
+master.name = "master";
+master.node_id = +(await grid3.capacity.filterNodes(masterQueryOptions))[0].nodeId;
+master.cpu = 1;
+master.memory = 1024;
+master.rootfs_size = 0;
+master.disk_size = 1;
+master.public_ip = false;
+master.planetary = true;
+
+ // create k8s node Object
+const worker = new KubernetesNodeModel();
+worker.name = "worker";
+worker.node_id = +(await grid3.capacity.filterNodes(workerQueryOptions))[0].nodeId;
+worker.cpu = 1;
+worker.memory = 1024;
+worker.rootfs_size = 0;
+worker.disk_size = 1;
+worker.public_ip = false;
+worker.planetary = true;
+
+```
+
+### Building cluster
+
+Here we specify the cluster project name, cluster secret, network model to be used, master and workers nodes and sshkey to access them
+
+```ts
+// create k8s Object
+const k = new K8SModel();
+k.name = "testk8s";
+k.secret = "secret";
+k.network = n;
+k.masters = [master];
+k.workers = [worker];
+k.metadata = "{'testk8s': true}";
+k.description = "test deploying k8s via ts grid3 client";
+k.ssh_key = config.ssh_key;
+```
+
+### Deploying
+
+use `deploy` function to deploy the kubernetes project
+
+```ts
+const res = await grid3.k8s.deploy(k);
+log(res);
+```
+
+### Getting deployment information
+
+```ts
+const l = await grid3.k8s.getObj(k.name);
+log(l);
+```
+
+### Deleting deployment
+
+```ts
+const d = await grid3.k8s.delete({ name: k.name });
+log(d);
+```
diff --git a/collections/documentation/developers/javascript/grid3_javascript_kvstore.md b/collections/documentation/developers/javascript/grid3_javascript_kvstore.md
new file mode 100644
index 0000000..5075086
--- /dev/null
+++ b/collections/documentation/developers/javascript/grid3_javascript_kvstore.md
@@ -0,0 +1,101 @@
+Using TFChain KVStore
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Prerequisites](#prerequisites)
+- [Example code](#example-code)
+ - [setting values](#setting-values)
+ - [getting key](#getting-key)
+ - [listing keys](#listing-keys)
+ - [deleting key](#deleting-key)
+
+***
+
+## Introduction
+
+As part of the tfchain, we support a keyvalue store module that can be used for any value within `2KB` range. practically it's used to save the user configurations state, so it can be built up again on any machine, given they used the same mnemonics and same secret.
+
+## Prerequisites
+
+- Make sure you have your [client](./grid3_javascript_loadclient.md) prepared
+
+## Example code
+
+```ts
+import { getClient } from "./client_loader";
+import { log } from "./utils";
+
+/*
+KVStore example usage:
+*/
+async function main() {
+ //For creating grid3 client with KVStore, you need to specify the KVStore storage type in the pram:
+
+ const gridClient = await getClient();
+
+ //then every module will use the KVStore to save its configuration and restore it.
+
+ // also you can use it like this:
+ const db = gridClient.kvstore;
+
+ // set key
+ const key = "hamada";
+ const exampleObj = {
+ key1: "value1",
+ key2: 2,
+ };
+ // set key
+ await db.set({ key, value: JSON.stringify(exampleObj) });
+
+ // list all the keys
+ const keys = await db.list();
+ log(keys);
+
+ // get the key
+ const data = await db.get({ key });
+ log(JSON.parse(data));
+
+ // remove the key
+ await db.remove({ key });
+
+ await gridClient.disconnect();
+}
+
+main();
+
+```
+
+### setting values
+
+`db.set` is used to set key to any value `serialized as string`
+
+```ts
+await db.set({ key, value: JSON.stringify(exampleObj) });
+```
+
+### getting key
+
+`db.get` is used to get a specific key
+
+```ts
+const data = await db.get({ key });
+log(JSON.parse(data));
+```
+
+### listing keys
+
+`db.list` is used to list all the keys.
+
+```ts
+const keys = await db.list();
+log(keys);
+```
+
+### deleting key
+
+`db.remove` is used to delete a specific key.
+
+```ts
+await db.remove({ key });
+```
diff --git a/collections/documentation/developers/javascript/grid3_javascript_loadclient.md b/collections/documentation/developers/javascript/grid3_javascript_loadclient.md
new file mode 100644
index 0000000..fc7c025
--- /dev/null
+++ b/collections/documentation/developers/javascript/grid3_javascript_loadclient.md
@@ -0,0 +1,68 @@
+ Grid3 Client
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Client Configurations](#client-configurations)
+- [Creating/Initializing The Grid3 Client](#creatinginitializing-the-grid3-client)
+- [What is `rmb-rs` | Reliable Message Bus --rust](#what-is-rmb-rs--reliable-message-bus---rust)
+- [Grid3 Client Options](#grid3-client-options)
+
+## Introduction
+
+Grid3 Client is a client used for deploying workloads (VMs, ZDBs, k8s, etc.) on the TFGrid.
+
+## Client Configurations
+
+so you have to set up your configuration file to be like this:
+
+```json
+{
+ "network": "dev",
+ "mnemonic": "",
+ "storeSecret": "secret",
+ "ssh_key": ""
+}
+```
+
+## Creating/Initializing The Grid3 Client
+
+```ts
+async function getClient(): Promise {
+ const gridClient = new GridClient({
+ network: "dev", // can be dev, qa, test, main, or custom
+ mnemonic: "",
+ });
+ await gridClient.connect();
+
+ return gridClient;
+ }
+```
+
+The grid client uses `rmb-rs` tool to send requests to/from nodes.
+
+## What is `rmb-rs` | Reliable Message Bus --rust
+
+Reliable message bus is a secure communication panel that allows bots to communicate together in a chat like way. It makes it very easy to host a service or a set of functions to be used by anyone, even if your service is running behind NAT.
+
+Out of the box RMB provides the following:
+
+- Guarantee authenticity of the messages. You are always sure that the received message is from whoever is pretending to be
+- End to End encryption
+- Support for 3rd party hosted relays. Anyone can host a relay and people can use it safely since there is no way messages can be inspected while
+using e2e. That's similar to home servers by matrix
+
+## Grid3 Client Options
+
+- network: `dev` for devnet, `test` for testnet
+- mnemonics: used for signing the requests.
+- storeSecret: used to encrypt data while storing in backend. It's any word that will be used for encrypting/decrypting the keys on threefold key-value store. If left empty, the Grid client will use the mnemonics as the storeSecret.
+- BackendStorage : can be `auto` which willl automatically adapt if running in node environment to use `filesystem backend` or the browser enviornment to use `localstorage backend`. Also you can set it to `kvstore` to use the tfchain keyvalue store module.
+- keypairType: is defaulted to `sr25519`, most likely you will never need to change it. `ed25519` is supported too.
+
+for more details, check [client options](https://github.com/threefoldtech/tfgrid-sdk-ts/blob/development/packages/grid_client/docs/client_configuration.md)
+
+> Note: The choice of the node is completely up to the user at this point. They need to do the capacity planning. Check [Node Finder](../../dashboard/deploy/node_finder.md) to know which nodes fits your deployment criteria.
+
+Check the document for [capacity planning using code](../javascript/grid3_javascript_capacity_planning.md) if you want to automate it
+> Note: this feature is still experimental
diff --git a/collections/documentation/developers/javascript/grid3_javascript_qsfs.md b/collections/documentation/developers/javascript/grid3_javascript_qsfs.md
new file mode 100644
index 0000000..df6b02a
--- /dev/null
+++ b/collections/documentation/developers/javascript/grid3_javascript_qsfs.md
@@ -0,0 +1,297 @@
+Deploying a VM with QSFS
+
+Table of Contents
+
+- [Prerequisites](#prerequisites)
+- [Code Example](#code-example)
+- [Detailed Explanation](#detailed-explanation)
+ - [Getting the Client](#getting-the-client)
+ - [Preparing QSFS](#preparing-qsfs)
+ - [Deploying a VM with QSFS](#deploying-a-vm-with-qsfs)
+ - [Getting the Deployment Information](#getting-the-deployment-information)
+ - [Deleting a Deployment](#deleting-a-deployment)
+
+***
+
+## Prerequisites
+
+First, make sure that you have your [client](./grid3_javascript_loadclient.md) prepared.
+
+## Code Example
+
+```ts
+import { FilterOptions, MachinesModel, QSFSZDBSModel } from "../src";
+import { config, getClient } from "./client_loader";
+import { log } from "./utils";
+
+async function main() {
+ const grid3 = await getClient();
+
+ const qsfs_name = "wed2710q1";
+ const machines_name = "wed2710t1";
+
+ const vmQueryOptions: FilterOptions = {
+ cru: 1,
+ mru: 1, // GB
+ sru: 1,
+ availableFor: grid3.twinId,
+ farmId: 1,
+ };
+
+ const qsfsQueryOptions: FilterOptions = {
+ hru: 6,
+ availableFor: grid3.twinId,
+ farmId: 1,
+ };
+
+ const qsfsNodes = [];
+
+ const allNodes = await grid3.capacity.filterNodes(qsfsQueryOptions);
+ if (allNodes.length >= 2) {
+ qsfsNodes.push(+allNodes[0].nodeId, +allNodes[1].nodeId);
+ } else {
+ throw Error("Couldn't find nodes for qsfs");
+ }
+
+ const vmNode = +(await grid3.capacity.filterNodes(vmQueryOptions))[0].nodeId;
+
+ const qsfs: QSFSZDBSModel = {
+ name: qsfs_name,
+ count: 8,
+ node_ids: qsfsNodes,
+ password: "mypassword",
+ disk_size: 1,
+ description: "my qsfs test",
+ metadata: "",
+ };
+
+ const vms: MachinesModel = {
+ name: machines_name,
+ network: {
+ name: "wed2710n1",
+ ip_range: "10.201.0.0/16",
+ },
+ machines: [
+ {
+ name: "wed2710v1",
+ node_id: vmNode,
+ disks: [
+ {
+ name: "wed2710d1",
+ size: 1,
+ mountpoint: "/mydisk",
+ },
+ ],
+ qsfs_disks: [
+ {
+ qsfs_zdbs_name: qsfs_name,
+ name: "wed2710d2",
+ minimal_shards: 2,
+ expected_shards: 4,
+ encryption_key: "hamada",
+ prefix: "hamada",
+ cache: 1,
+ mountpoint: "/myqsfsdisk",
+ },
+ ],
+ public_ip: false,
+ public_ip6: false,
+ planetary: true,
+ cpu: 1,
+ memory: 1024,
+ rootfs_size: 0,
+ flist: "https://hub.grid.tf/tf-official-apps/base:latest.flist",
+ entrypoint: "/sbin/zinit init",
+ env: {
+ SSH_KEY: config.ssh_key,
+ },
+ },
+ ],
+ metadata: "{'testVMs': true}",
+ description: "test deploying VMs via ts grid3 client",
+ };
+
+ async function cancel(grid3) {
+ // delete
+ const d = await grid3.machines.delete({ name: machines_name });
+ log(d);
+ const r = await grid3.qsfs_zdbs.delete({ name: qsfs_name });
+ log(r);
+ }
+ //deploy qsfs
+ const res = await grid3.qsfs_zdbs.deploy(qsfs);
+ log(">>>>>>>>>>>>>>>QSFS backend has been created<<<<<<<<<<<<<<<");
+ log(res);
+
+ const vm_res = await grid3.machines.deploy(vms);
+ log(">>>>>>>>>>>>>>>vm has been created<<<<<<<<<<<<<<<");
+ log(vm_res);
+
+ // get the deployment
+ const l = await grid3.machines.getObj(vms.name);
+ log(">>>>>>>>>>>>>>>Deployment result<<<<<<<<<<<<<<<");
+ log(l);
+
+ // await cancel(grid3);
+
+ await grid3.disconnect();
+}
+
+main();
+```
+
+## Detailed Explanation
+
+We present a detailed explanation of the example shown above.
+
+### Getting the Client
+
+```ts
+const grid3 = getClient();
+```
+
+### Preparing QSFS
+
+```ts
+const qsfs_name = "wed2710q1";
+const machines_name = "wed2710t1";
+```
+
+We prepare here some names to use across the client for the QSFS and the machines project
+
+```ts
+ const qsfsQueryOptions: FilterOptions = {
+ hru: 6,
+ availableFor: grid3.twinId,
+ farmId: 1,
+ };
+ const qsfsNodes = [];
+
+ const allNodes = await grid3.capacity.filterNodes(qsfsQueryOptions);
+ if (allNodes.length >= 2) {
+ qsfsNodes.push(+allNodes[0].nodeId, +allNodes[1].nodeId);
+ } else {
+ throw Error("Couldn't find nodes for qsfs");
+ }
+
+ const vmNode = +(await grid3.capacity.filterNodes(vmQueryOptions))[0].nodeId;
+
+ const qsfs: QSFSZDBSModel = {
+ name: qsfs_name,
+ count: 8,
+ node_ids: qsfsNodes,
+ password: "mypassword",
+ disk_size: 1,
+ description: "my qsfs test",
+ metadata: "",
+ };
+
+const res = await grid3.qsfs_zdbs.deploy(qsfs);
+log(">>>>>>>>>>>>>>>QSFS backend has been created<<<<<<<<<<<<<<<");
+log(res);
+```
+
+Here we deploy `8` ZDBs on nodes `2,3` with password `mypassword`, all of them having disk size of `10GB`
+
+### Deploying a VM with QSFS
+
+```ts
+const vmQueryOptions: FilterOptions = {
+ cru: 1,
+ mru: 1, // GB
+ sru: 1,
+ availableFor: grid3.twinId,
+ farmId: 1,
+};
+
+const vmNode = +(await grid3.capacity.filterNodes(vmQueryOptions))[0].nodeId;
+
+ // deploy vms
+const vms: MachinesModel = {
+ name: machines_name,
+ network: {
+ name: "wed2710n1",
+ ip_range: "10.201.0.0/16",
+ },
+ machines: [
+ {
+ name: "wed2710v1",
+ node_id: vmNode,
+ disks: [
+ {
+ name: "wed2710d1",
+ size: 1,
+ mountpoint: "/mydisk",
+ },
+ ],
+ qsfs_disks: [
+ {
+ qsfs_zdbs_name: qsfs_name,
+ name: "wed2710d2",
+ minimal_shards: 2,
+ expected_shards: 4,
+ encryption_key: "hamada",
+ prefix: "hamada",
+ cache: 1,
+ mountpoint: "/myqsfsdisk",
+ },
+ ],
+ public_ip: false,
+ public_ip6: false,
+ planetary: true,
+ cpu: 1,
+ memory: 1024,
+ rootfs_size: 0,
+ flist: "https://hub.grid.tf/tf-official-apps/base:latest.flist",
+ entrypoint: "/sbin/zinit init",
+ env: {
+ SSH_KEY: config.ssh_key,
+ },
+ },
+ ],
+ metadata: "{'testVMs': true}",
+ description: "test deploying VMs via ts grid3 client",
+};
+const vm_res = await grid3.machines.deploy(vms);
+log(">>>>>>>>>>>>>>>vm has been created<<<<<<<<<<<<<<<");
+log(vm_res);
+```
+
+So this deployment is almost similiar to what we have in the [vm deployment section](./grid3_javascript_vm.md). We only have a new section `qsfs_disks`
+
+```ts
+ qsfs_disks: [{
+ qsfs_zdbs_name: qsfs_name,
+ name: "wed2710d2",
+ minimal_shards: 2,
+ expected_shards: 4,
+ encryption_key: "hamada",
+ prefix: "hamada",
+ cache: 1,
+ mountpoint: "/myqsfsdisk"
+ }],
+```
+
+`qsfs_disks` is a list, representing all of the QSFS disks used within that VM.
+
+- `qsfs_zdbs_name`: that's the backend ZDBs we defined in the beginning
+- `expected_shards`: how many ZDBs that QSFS should be working with
+- `minimal_shards`: the minimal possible amount of ZDBs to recover the data with when losing disks e.g due to failure
+- `mountpoint`: where it will be mounted on the VM `/myqsfsdisk`
+
+### Getting the Deployment Information
+
+```ts
+const l = await grid3.machines.getObj(vms.name);
+log(l);
+```
+
+### Deleting a Deployment
+
+```ts
+// delete
+const d = await grid3.machines.delete({ name: machines_name });
+log(d);
+const r = await grid3.qsfs_zdbs.delete({ name: qsfs_name });
+log(r);
+```
diff --git a/collections/documentation/developers/javascript/grid3_javascript_qsfs_zdbs.md b/collections/documentation/developers/javascript/grid3_javascript_qsfs_zdbs.md
new file mode 100644
index 0000000..9c3a3f2
--- /dev/null
+++ b/collections/documentation/developers/javascript/grid3_javascript_qsfs_zdbs.md
@@ -0,0 +1,142 @@
+Deploying ZDBs for QSFS
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Prerequisites](#prerequisites)
+- [Example code](#example-code)
+- [Detailed explanation](#detailed-explanation)
+ - [Getting the client](#getting-the-client)
+ - [Preparing the nodes](#preparing-the-nodes)
+ - [Preparing ZDBs](#preparing-zdbs)
+ - [Deploying the ZDBs](#deploying-the-zdbs)
+ - [Getting deployment information](#getting-deployment-information)
+ - [Deleting a deployment](#deleting-a-deployment)
+
+***
+
+## Introduction
+
+We show how to deploy ZDBs for QSFS on the TFGrid with the Javascript client.
+
+## Prerequisites
+
+- Make sure you have your [client](./grid3_javascript_loadclient.md) prepared
+
+## Example code
+
+````typescript
+import { FilterOptions, QSFSZDBSModel } from "../src";
+import { getClient } from "./client_loader";
+import { log } from "./utils";
+
+async function main() {
+ const grid3 = await getClient();
+ const qsfs_name = "zdbsQsfsDemo";
+ const qsfsQueryOptions: FilterOptions = {
+ hru: 8,
+ availableFor: grid3.twinId,
+ farmId: 1,
+ };
+ const qsfsNodes = [];
+
+ const allNodes = await grid3.capacity.filterNodes(qsfsQueryOptions);
+
+ if (allNodes.length >= 2) {
+ qsfsNodes.push(+allNodes[0].nodeId, +allNodes[1].nodeId);
+ } else {
+ throw Error("Couldn't find nodes for qsfs");
+ }
+
+ const qsfs: QSFSZDBSModel = {
+ name: qsfs_name,
+ count: 12,
+ node_ids: qsfsNodes,
+ password: "mypassword",
+ disk_size: 1,
+ description: "my zdbs test",
+ metadata: "",
+ };
+ const deploy_res = await grid3.qsfs_zdbs.deploy(qsfs);
+ log(deploy_res);
+
+ const zdbs_data = await grid3.qsfs_zdbs.get({ name: qsfs_name });
+ log(zdbs_data);
+
+
+ await grid3.disconnect();
+}
+main();
+
+````
+
+## Detailed explanation
+
+### Getting the client
+
+```typescript
+const grid3 = getClient();
+```
+
+### Preparing the nodes
+
+we need to deploy the zdbs on two different nodes so, we setup the filters here to retrieve the available nodes.
+
+````typescript
+const qsfsQueryOptions: FilterOptions = {
+ hru: 16,
+ availableFor: grid3.twinId,
+ farmId: 1,
+};
+const qsfsNodes = [];
+
+const allNodes = await grid3.capacity.filterNodes(qsfsQueryOptions);
+
+if (allNodes.length >= 2) {
+ qsfsNodes.push(+allNodes[0].nodeId, +allNodes[1].nodeId);
+} else {
+ throw Error("Couldn't find nodes for qsfs");
+}
+````
+
+Now we have two nodes in `qsfsNode`.
+
+### Preparing ZDBs
+
+````typescript
+const qsfs_name = "zdbsQsfsDemo";
+````
+
+We prepare here a name to use across the client for the QSFS ZDBs
+
+### Deploying the ZDBs
+
+````typescript
+const qsfs: QSFSZDBSModel = {
+ name: qsfs_name,
+ count: 12,
+ node_ids: qsfsNodes,
+ password: "mypassword",
+ disk_size: 1,
+ description: "my qsfs test",
+ metadata: "",
+ };
+const deploy_res = await grid3.qsfs_zdbs.deploy(qsfs);
+log(deploy_res);
+````
+
+Here we deploy `12` ZDBs on nodes in `qsfsNode` with password `mypassword`, all of them having disk size of `1GB`, the client already add 4 zdbs for metadata.
+
+### Getting deployment information
+
+````typescript
+const zdbs_data = await grid3.qsfs_zdbs.get({ name: qsfs_name });
+log(zdbs_data);
+````
+
+### Deleting a deployment
+
+````typescript
+const delete_response = await grid3.qsfs_zdbs.delete({ name: qsfs_name });
+log(delete_response);
+````
diff --git a/collections/documentation/developers/javascript/grid3_javascript_readme.md b/collections/documentation/developers/javascript/grid3_javascript_readme.md
new file mode 100644
index 0000000..7072d19
--- /dev/null
+++ b/collections/documentation/developers/javascript/grid3_javascript_readme.md
@@ -0,0 +1,24 @@
+ Javascript Client
+
+This section covers developing projects on top of Threefold Grid using Javascript language.
+
+Javascript has a huge ecosystem, and first class citizen when it comes to blockchain technologies like substrate and that was one of the reasons for it to become one the very first supported languages on the grid.
+
+Please make sure to check the [basics](../../system_administrators/getstarted/tfgrid3_getstarted.md) before continuing.
+
+ Table of Contents
+
+- [Installation](./grid3_javascript_installation.md)
+- [Loading Client](./grid3_javascript_loadclient.md)
+- [Deploy a VM](./grid3_javascript_vm.md)
+- [Capacity Planning](./grid3_javascript_capacity_planning.md)
+- [Deploy Multiple VMs](./grid3_javascript_vms.md)
+- [Deploy CapRover](./grid3_javascript_caprover.md)
+- [Gateways](./grid3_javascript_vm_gateways.md)
+- [Deploy a Kubernetes Cluster](./grid3_javascript_kubernetes.md)
+- [Deploy a ZDB](./grid3_javascript_zdb.md)
+- [Deploy ZDBs for QSFS](./grid3_javascript_qsfs_zdbs.md)
+- [QSFS](./grid3_javascript_qsfs.md)
+- [Key Value Store](./grid3_javascript_kvstore.md)
+- [VM with Wireguard and Gateway](./grid3_wireguard_gateway.md)
+- [GPU Support](./grid3_javascript_gpu_support.md)
\ No newline at end of file
diff --git a/collections/documentation/developers/javascript/grid3_javascript_run_scripts.md b/collections/documentation/developers/javascript/grid3_javascript_run_scripts.md
new file mode 100644
index 0000000..7c3a2d5
--- /dev/null
+++ b/collections/documentation/developers/javascript/grid3_javascript_run_scripts.md
@@ -0,0 +1,15 @@
+## How to run the scripts
+
+- Set your grid3 client configuration in `scripts/client_loader.ts` or easily use one of `config.json`
+- update your customized deployments specs
+- Run using [ts-node](https://www.npmjs.com/ts-node)
+
+```bash
+npx ts-node --project tsconfig-node.json scripts/zdb.ts
+```
+
+or
+
+```bash
+yarn run ts-node --project tsconfig-node.json scripts/zdb.ts
+```
diff --git a/collections/documentation/developers/javascript/grid3_javascript_vm.md b/collections/documentation/developers/javascript/grid3_javascript_vm.md
new file mode 100644
index 0000000..5c9069e
--- /dev/null
+++ b/collections/documentation/developers/javascript/grid3_javascript_vm.md
@@ -0,0 +1,194 @@
+
+ Deploying a VM
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Example](#example)
+- [Detailed Explanation](#detailed-explanation)
+ - [Building Network](#building-network)
+- [Building the Disk Model](#building-the-disk-model)
+- [Building the VM](#building-the-vm)
+- [Building VMs Collection](#building-vms-collection)
+- [deployment](#deployment)
+- [Getting Deployment Information](#getting-deployment-information)
+- [Deleting a Deployment](#deleting-a-deployment)
+
+***
+
+## Introduction
+
+We present information on how to deploy a VM with the Javascript client with concrete examples.
+
+## Example
+
+```ts
+import { DiskModel, FilterOptions, MachineModel, MachinesModel, NetworkModel } from "../src";
+import { config, getClient } from "./client_loader";
+import { log } from "./utils";
+
+async function main() {
+ const grid3 = await getClient();
+
+ // create network Object
+ const n = new NetworkModel();
+ n.name = "dynamictest";
+ n.ip_range = "10.249.0.0/16";
+
+ // create disk Object
+ const disk = new DiskModel();
+ disk.name = "dynamicDisk";
+ disk.size = 8;
+ disk.mountpoint = "/testdisk";
+
+ const vmQueryOptions: FilterOptions = {
+ cru: 1,
+ mru: 1, // GB
+ sru: 1,
+ availableFor: grid3.twinId,
+ country: "Belgium",
+ };
+
+ // create vm node Object
+ const vm = new MachineModel();
+ vm.name = "testvm";
+ vm.node_id = +(await grid3.capacity.filterNodes(vmQueryOptions))[0].nodeId; // TODO: allow random choice
+ vm.disks = [disk];
+ vm.public_ip = false;
+ vm.planetary = true;
+ vm.cpu = 1;
+ vm.memory = 1024;
+ vm.rootfs_size = 0;
+ vm.flist = "https://hub.grid.tf/tf-official-apps/base:latest.flist";
+ vm.entrypoint = "/sbin/zinit init";
+ vm.env = {
+ SSH_KEY: config.ssh_key,
+ };
+
+ // create VMs Object
+ const vms = new MachinesModel();
+ vms.name = "dynamicVMS";
+ vms.network = n;
+ vms.machines = [vm];
+ vms.metadata = "{'testVMs': true}";
+ vms.description = "test deploying VMs via ts grid3 client";
+
+ // deploy vms
+ const res = await grid3.machines.deploy(vms);
+ log(res);
+
+ // get the deployment
+ const l = await grid3.machines.getObj(vms.name);
+ log(l);
+
+ // // delete
+ // const d = await grid3.machines.delete({ name: vms.name });
+ // log(d);
+
+ await grid3.disconnect();
+}
+
+main();
+```
+
+## Detailed Explanation
+
+### Building Network
+
+```ts
+// create network Object
+const n = new NetworkModel();
+n.name = "dynamictest";
+n.ip_range = "10.249.0.0/16";
+```
+
+Here we prepare the network model that is going to be used by specifying a name to our network and the range it will be spanning over
+
+## Building the Disk Model
+
+```ts
+// create disk Object
+const disk = new DiskModel();
+disk.name = "dynamicDisk";
+disk.size = 8;
+disk.mountpoint = "/testdisk";
+```
+
+here we create the disk model specifying its name, size in GB and where it will be mounted eventually
+
+## Building the VM
+
+```ts
+// create vm node Object
+const vm = new MachineModel();
+vm.name = "testvm";
+vm.node_id = +(await grid3.capacity.filterNodes(vmQueryOptions))[0].nodeId; // TODO: allow random choice
+vm.disks = [disk];
+vm.public_ip = false;
+vm.planetary = true;
+vm.cpu = 1;
+vm.memory = 1024;
+vm.rootfs_size = 0;
+vm.flist = "https://hub.grid.tf/tf-official-apps/base:latest.flist";
+vm.entrypoint = "/sbin/zinit init";
+vm.env = {
+ SSH_KEY: config.ssh_key,
+};
+```
+
+Now we go to the VM model, that will be used to build our `zmachine` object
+
+We need to specify its
+
+- name
+- node_id: where it will get deployed
+- disks: disks model collection
+- memory
+- root filesystem size
+- flist: the image it is going to start from. Check the [supported flists](../flist/grid3_supported_flists.md)
+- entry point: entrypoint command / script to execute
+- env: has the environment variables needed e.g sshkeys used
+- public ip: if we want to have a public ip attached to the VM
+- planetary: to enable planetary network on VM
+
+## Building VMs Collection
+
+```ts
+// create VMs Object
+const vms = new MachinesModel();
+vms.name = "dynamicVMS";
+vms.network = n;
+vms.machines = [vm];
+vms.metadata = "{'testVMs': true}";
+vms.description = "test deploying VMs via ts grid3 client";
+```
+
+Here it's quite simple we can add one or more VM to the `machines` property to have them deployed as part of our project
+
+## deployment
+
+```ts
+// deploy vms
+const res = await grid3.machines.deploy(vms);
+log(res);
+```
+
+## Getting Deployment Information
+
+can do so based on the name you gave to the `vms` collection
+
+```ts
+// get the deployment
+const l = await grid3.machines.getObj(vms.name);
+log(l);
+```
+
+## Deleting a Deployment
+
+```ts
+// delete
+const d = await grid3.machines.delete({ name: vms.name });
+log(d);
+```
+
+In the underlying layer we cancel the contracts that were created on the chain and as a result all of the workloads tied to his project will get deleted.
diff --git a/collections/documentation/developers/javascript/grid3_javascript_vm_gateways.md b/collections/documentation/developers/javascript/grid3_javascript_vm_gateways.md
new file mode 100644
index 0000000..052d9f3
--- /dev/null
+++ b/collections/documentation/developers/javascript/grid3_javascript_vm_gateways.md
@@ -0,0 +1,189 @@
+ Deploying a VM and exposing it over a Gateway Prefix
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Example code](#example-code)
+- [Detailed explanation](#detailed-explanation)
+ - [deploying](#deploying)
+ - [getting deployment object](#getting-deployment-object)
+ - [deletion](#deletion)
+- [Deploying a VM and exposing it over a Gateway using a Full domain](#deploying-a-vm-and-exposing-it-over-a-gateway-using-a-full-domain)
+- [Example code](#example-code-1)
+- [Detailed explanation](#detailed-explanation-1)
+ - [deploying](#deploying-1)
+ - [get deployment object](#get-deployment-object)
+ - [deletion](#deletion-1)
+
+***
+
+## Introduction
+
+After the [deployment of a VM](./grid3_javascript_vm.md), now it's time to expose it to the world
+
+## Example code
+
+```ts
+import { FilterOptions, GatewayNameModel } from "../src";
+import { getClient } from "./client_loader";
+import { log } from "./utils";
+
+// read more about the gateway types in this doc: https://github.com/threefoldtech/zos/tree/main/docs/gateway
+async function main() {
+ const grid3 = await getClient();
+
+ const gatewayQueryOptions: FilterOptions = {
+ gateway: true,
+ farmId: 1,
+ };
+
+ const gw = new GatewayNameModel();
+ gw.name = "test";
+ gw.node_id = +(await grid3.capacity.filterNodes(gatewayQueryOptions))[0].nodeId;
+ gw.tls_passthrough = false;
+ // the backends have to be in this format `http://ip:port` or `https://ip:port`, and the `ip` pingable from the node so using the ygg ip or public ip if available.
+ gw.backends = ["http://185.206.122.35:8000"];
+
+ // deploy
+ const res = await grid3.gateway.deploy_name(gw);
+ log(res);
+
+ // get the deployment
+ const l = await grid3.gateway.getObj(gw.name);
+ log(l);
+
+ // // delete
+ // const d = await grid3.gateway.delete_name({ name: gw.name });
+ // log(d);
+
+ grid3.disconnect();
+}
+
+main();
+
+```
+
+## Detailed explanation
+
+```ts
+const gw = new GatewayNameModel();
+gw.name = "test";
+gw.node_id = +(await grid3.capacity.filterNodes(gatewayQueryOptions))[0].nodeId;
+gw.tls_passthrough = false;
+gw.backends = ["http://185.206.122.35:8000"];
+```
+
+- we created a gateway name model and gave it a `name` -that's why it's called GatewayName- `test` to be deployed on gateway node to end up with a domain `test.gent01.devnet.grid.tf`,
+- we create a proxy for the gateway to send the traffic coming to `test.ghent01.devnet.grid.tf` to the backend `http://185.206.122.35`, we say `tls_passthrough is false` to let the gateway terminate the traffic, if you replace it with `true` your backend service needs to be able to do the TLS termination
+
+### deploying
+
+```ts
+// deploy
+const res = await grid3.gateway.deploy_name(gw);
+log(res);
+```
+
+this deploys `GatewayName` on the grid
+
+### getting deployment object
+
+```ts
+const l = await grid3.gateway.getObj(gw.name);
+log(l);
+```
+
+getting the deployment information can be done using `getObj`
+
+### deletion
+
+```ts
+const d = await grid3.gateway.delete_name({ name: gw.name });
+log(d);
+```
+
+## Deploying a VM and exposing it over a Gateway using a Full domain
+
+After the [deployment of a VM](./grid3_javascript_vm.md), now it's time to expose it to the world
+
+## Example code
+
+```ts
+import { FilterOptions, GatewayFQDNModel } from "../src";
+import { getClient } from "./client_loader";
+import { log } from "./utils";
+
+// read more about the gateway types in this doc: https://github.com/threefoldtech/zos/tree/main/docs/gateway
+async function main() {
+ const grid3 = await getClient();
+
+ const gatewayQueryOptions: FilterOptions = {
+ gateway: true,
+ farmId: 1,
+ };
+ const gw = new GatewayFQDNModel();
+ gw.name = "applyFQDN";
+ gw.node_id = +(await grid3.capacity.filterNodes(gatewayQueryOptions))[0].nodeId;
+ gw.fqdn = "test.hamada.grid.tf";
+ gw.tls_passthrough = false;
+ // the backends have to be in this format `http://ip:port` or `https://ip:port`, and the `ip` pingable from the node so using the ygg ip or public ip if available.
+ gw.backends = ["http://185.206.122.35:8000"];
+
+ // deploy
+ const res = await grid3.gateway.deploy_fqdn(gw);
+ log(res);
+
+ // get the deployment
+ const l = await grid3.gateway.getObj(gw.name);
+ log(l);
+
+ // // delete
+ // const d = await grid3.gateway.delete_fqdn({ name: gw.name });
+ // log(d);
+
+ grid3.disconnect();
+}
+
+main();
+```
+
+## Detailed explanation
+
+```ts
+const gw = new GatewayFQDNModel();
+gw.name = "applyFQDN";
+gw.node_id = 1;
+gw.fqdn = "test.hamada.grid.tf";
+gw.tls_passthrough = false;
+gw.backends = ["my yggdrasil IP"];
+```
+
+- we created a `GatewayFQDNModel` and gave it a name `applyFQDNN` to be deployed on gateway node `1` and specified the fully qualified domain `fqdn` to a domain we own `test.hamada.grid.tf`
+- we created a record on our name provider for `test.hamada.grid.tf` to point to the IP of gateway node `1`
+- we specified the backened would be an yggdrassil ip so once this is deployed when we go to `test.hamada.grid.tf` we go to the gateway server and from their our traffic goes to the backend.
+
+### deploying
+
+```ts
+// deploy
+const res = await grid3.gateway.deploy_fqdn(gw);
+log(res);
+```
+
+this deploys `GatewayName` on the grid
+
+### get deployment object
+
+```ts
+const l = await grid3.gateway.getObj(gw.name);
+log(l);
+```
+
+getting the deployment information can be done using `getObj`
+
+### deletion
+
+```ts
+const d = await grid3.gateway.delete_fqdn({ name: gw.name });
+log(d);
+```
diff --git a/collections/documentation/developers/javascript/grid3_javascript_vms.md b/collections/documentation/developers/javascript/grid3_javascript_vms.md
new file mode 100644
index 0000000..b928007
--- /dev/null
+++ b/collections/documentation/developers/javascript/grid3_javascript_vms.md
@@ -0,0 +1,108 @@
+
+ Deploying multiple VMs
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Example code](#example-code)
+
+***
+
+## Introduction
+
+It is possible to deploy multiple VMs with the Javascript client.
+
+## Example code
+
+```ts
+import { DiskModel, FilterOptions, MachineModel, MachinesModel, NetworkModel } from "../src";
+import { config, getClient } from "./client_loader";
+import { log } from "./utils";
+
+async function main() {
+ const grid3 = await getClient();
+
+ // create network Object
+ const n = new NetworkModel();
+ n.name = "monNetwork";
+ n.ip_range = "10.238.0.0/16";
+
+ // create disk Object
+ const disk1 = new DiskModel();
+ disk1.name = "newDisk1";
+ disk1.size = 1;
+ disk1.mountpoint = "/newDisk1";
+
+ const vmQueryOptions: FilterOptions = {
+ cru: 1,
+ mru: 1, // GB
+ sru: 1,
+ availableFor: grid3.twinId,
+ farmId: 1,
+ };
+
+ // create vm node Object
+ const vm1 = new MachineModel();
+ vm1.name = "testvm1";
+ vm1.node_id = +(await grid3.capacity.filterNodes(vmQueryOptions))[0].nodeId;
+ vm1.disks = [disk1];
+ vm1.public_ip = false;
+ vm1.planetary = true;
+ vm1.cpu = 1;
+ vm1.memory = 1024;
+ vm1.rootfs_size = 0;
+ vm1.flist = "https://hub.grid.tf/tf-official-apps/base:latest.flist";
+ vm1.entrypoint = "/sbin/zinit init";
+ vm1.env = {
+ SSH_KEY: config.ssh_key,
+ };
+
+ // create disk Object
+ const disk2 = new DiskModel();
+ disk2.name = "newDisk2";
+ disk2.size = 1;
+ disk2.mountpoint = "/newDisk2";
+
+ // create another vm node Object
+ const vm2 = new MachineModel();
+ vm2.name = "testvm2";
+ vm2.node_id = +(await grid3.capacity.filterNodes(vmQueryOptions))[1].nodeId;
+ vm2.disks = [disk2];
+ vm2.public_ip = false;
+ vm2.planetary = true;
+ vm2.cpu = 1;
+ vm2.memory = 1024;
+ vm2.rootfs_size = 0;
+ vm2.flist = "https://hub.grid.tf/tf-official-apps/base:latest.flist";
+ vm2.entrypoint = "/sbin/zinit init";
+ vm2.env = {
+ SSH_KEY: config.ssh_key,
+ };
+
+ // create VMs Object
+ const vms = new MachinesModel();
+ vms.name = "monVMS";
+ vms.network = n;
+ vms.machines = [vm1, vm2];
+ vms.metadata = "{'testVMs': true}";
+ vms.description = "test deploying VMs via ts grid3 client";
+
+ // deploy vms
+ const res = await grid3.machines.deploy(vms);
+ log(res);
+
+ // get the deployment
+ const l = await grid3.machines.getObj(vms.name);
+ log(l);
+
+ // // delete
+ // const d = await grid3.machines.delete({ name: vms.name });
+ // log(d);
+
+ await grid3.disconnect();
+}
+
+main();
+```
+
+It's similiar to the previous section of [deploying a single VM](../javascript/grid3_javascript_vm.md), but just adds more vm objects to vms collection.
diff --git a/collections/documentation/developers/javascript/grid3_javascript_zdb.md b/collections/documentation/developers/javascript/grid3_javascript_zdb.md
new file mode 100644
index 0000000..d773269
--- /dev/null
+++ b/collections/documentation/developers/javascript/grid3_javascript_zdb.md
@@ -0,0 +1,143 @@
+Deploying ZDB
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Prerequisites](#prerequisites)
+- [Example code](#example-code)
+- [Detailed explanation](#detailed-explanation)
+ - [Getting the client](#getting-the-client)
+ - [Building the model](#building-the-model)
+ - [preparing ZDBs collection](#preparing-zdbs-collection)
+ - [Deployment](#deployment)
+ - [Getting Deployment information](#getting-deployment-information)
+ - [Deleting a deployment](#deleting-a-deployment)
+
+***
+
+## Introduction
+
+We show how to deploy ZDB on the TFGrid with the Javascript client.
+
+## Prerequisites
+
+- Make sure you have your [client](./grid3_javascript_loadclient.md) prepared
+
+## Example code
+
+```ts
+import { FilterOptions, ZDBModel, ZdbModes, ZDBSModel } from "../src";
+import { getClient } from "./client_loader";
+import { log } from "./utils";
+
+async function main() {
+ const grid3 = await getClient();
+
+ const zdbQueryOptions: FilterOptions = {
+ sru: 1,
+ hru: 1,
+ availableFor: grid3.twinId,
+ farmId: 1,
+ };
+
+ // create zdb object
+ const zdb = new ZDBModel();
+ zdb.name = "hamada";
+ zdb.node_id = +(await grid3.capacity.filterNodes(zdbQueryOptions))[0].nodeId;
+ zdb.mode = ZdbModes.user;
+ zdb.disk_size = 1;
+ zdb.publicNamespace = false;
+ zdb.password = "testzdb";
+
+ // create zdbs object
+ const zdbs = new ZDBSModel();
+ zdbs.name = "tttzdbs";
+ zdbs.zdbs = [zdb];
+ zdbs.metadata = '{"test": "test"}';
+
+ // deploy zdb
+ const res = await grid3.zdbs.deploy(zdbs);
+ log(res);
+
+ // get the deployment
+ const l = await grid3.zdbs.getObj(zdbs.name);
+ log(l);
+
+ // // delete
+ // const d = await grid3.zdbs.delete({ name: zdbs.name });
+ // log(d);
+
+ await grid3.disconnect();
+}
+
+main();
+```
+
+## Detailed explanation
+
+### Getting the client
+
+```ts
+const grid3 = getClient();
+```
+
+### Building the model
+
+```ts
+// create zdb object
+const zdb = new ZDBModel();
+zdb.name = "hamada";
+zdb.node_id = +(await grid3.capacity.filterNodes(zdbQueryOptions))[0].nodeId;
+zdb.mode = ZdbModes.user;
+zdb.disk_size = 1;
+zdb.publicNamespace = false;
+zdb.password = "testzdb";
+```
+
+Here we define a `ZDB model` and setting the relevant properties e.g
+
+- name
+- node_id : where to deploy on
+- mode: `user` or `seq`
+- disk_size: disk size in GB
+- publicNamespace: a public namespace can be read-only if a password is set
+- password: namespace password
+
+### preparing ZDBs collection
+
+```ts
+// create zdbs object
+const zdbs = new ZDBSModel();
+zdbs.name = "tttzdbs";
+zdbs.zdbs = [zdb];
+zdbs.metadata = '{"test": "test"}';
+```
+
+you can attach multiple ZDBs into the collection and send it for deployment
+
+### Deployment
+
+```ts
+const res = await grid3.zdbs.deploy(zdbs);
+log(res);
+```
+
+### Getting Deployment information
+
+`getObj` gives detailed information about the workload.
+
+```ts
+// get the deployment
+const l = await grid3.zdbs.getObj(zdbs.name);
+log(l);
+```
+
+### Deleting a deployment
+
+`.delete` method helps cancelling the relevant contracts related to that ZDBs deployment
+
+```ts
+// delete
+const d = await grid3.zdbs.delete({ name: zdbs.name });
+log(d);
+```
diff --git a/collections/documentation/developers/javascript/grid3_wireguard_gateway.md b/collections/documentation/developers/javascript/grid3_wireguard_gateway.md
new file mode 100644
index 0000000..2cb1b89
--- /dev/null
+++ b/collections/documentation/developers/javascript/grid3_wireguard_gateway.md
@@ -0,0 +1,302 @@
+ Deploying a VM with Wireguard and Gateway
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Client Configurations](#client-configurations)
+- [Code Example](#code-example)
+- [Detailed Explanation](#detailed-explanation)
+ - [Get the Client](#get-the-client)
+ - [Get the Nodes](#get-the-nodes)
+ - [Deploy the VM](#deploy-the-vm)
+ - [Deploy the Gateway](#deploy-the-gateway)
+ - [Get the Deployments Information](#get-the-deployments-information)
+ - [Disconnect the Client](#disconnect-the-client)
+ - [Delete the Deployments](#delete-the-deployments)
+- [Conclusion](#conclusion)
+
+***
+
+## Introduction
+
+We present here the relevant information when it comes to deploying a virtual machine with Wireguard and a gateway.
+
+
+
+
+## Client Configurations
+
+To configure the client, have a look at [this section](./grid3_javascript_loadclient.md).
+
+
+
+## Code Example
+
+```ts
+import { FilterOptions, GatewayNameModel, GridClient, MachineModel, MachinesModel, NetworkModel } from "../src";
+import { config, getClient } from "./client_loader";
+import { log } from "./utils";
+
+function createNetworkModel(gwNode: number, name: string): NetworkModel {
+ return {
+ name,
+ addAccess: true,
+ accessNodeId: gwNode,
+ ip_range: "10.238.0.0/16",
+ } as NetworkModel;
+}
+function createMachineModel(node: number) {
+ return {
+ name: "testvm1",
+ node_id: node,
+ public_ip: false,
+ planetary: true,
+ cpu: 1,
+ memory: 1024 * 2,
+ rootfs_size: 0,
+ disks: [],
+ flist: "https://hub.grid.tf/tf-official-apps/threefoldtech-ubuntu-22.04.flist",
+ entrypoint: "/usr/bin/python3 -m http.server --bind ::",
+ env: {
+ SSH_KEY: config.ssh_key,
+ },
+ } as MachineModel;
+}
+function createMachinesModel(vm: MachineModel, network: NetworkModel): MachinesModel {
+ return {
+ name: "newVMs",
+ network,
+ machines: [vm],
+ metadata: "",
+ description: "test deploying VMs with wireguard via ts grid3 client",
+ } as MachinesModel;
+}
+function createGwModel(node_id: number, ip: string, networkName: string, name: string, port: number) {
+ return {
+ name,
+ node_id,
+ tls_passthrough: false,
+ backends: [`http://${ip}:${port}`],
+ network: networkName,
+ } as GatewayNameModel;
+}
+
+async function main() {
+ const grid3 = await getClient();
+
+ const gwNode = +(await grid3.capacity.filterNodes({ gateway: true }))[0].nodeId;
+
+ const vmQueryOptions: FilterOptions = {
+ cru: 1,
+ mru: 2, // GB
+ availableFor: grid3.twinId,
+ farmId: 1,
+ };
+ const vmNode = +(await grid3.capacity.filterNodes(vmQueryOptions))[0].nodeId;
+
+ const network = createNetworkModel(gwNode, "monNetwork");
+ const vm = createMachineModel(vmNode);
+ const machines = createMachinesModel(vm, network);
+ log(`Deploying vm on node: ${vmNode}, with network node: ${gwNode}`);
+
+ // deploy the vm
+ const vmResult = await grid3.machines.deploy(machines);
+ log(vmResult);
+
+ const deployedVm = await grid3.machines.getObj(machines.name);
+ log("+++ deployed vm +++");
+ log(deployedVm);
+
+ // deploy the gateway
+ const vmPrivateIP = (deployedVm as { interfaces: { ip: string }[] }[])[0].interfaces[0].ip;
+ const gateway = createGwModel(gwNode, vmPrivateIP, network.name, "pyserver", 8000);
+ log(`deploying gateway ${network.name} on node ${gwNode}`);
+
+ const gatewayResult = await grid3.gateway.deploy_name(gateway);
+ log(gatewayResult);
+
+ log("+++ Deployed gateway +++");
+
+ const deployedGw = await grid3.gateway.getObj(gateway.name);
+ log(deployedGw);
+
+ await grid3.disconnect();
+}
+
+main();
+
+```
+
+
+## Detailed Explanation
+
+What we need to do with that code is: Deploy a name gateway with the wireguard IP as the backend; that allows accessing a server inside the vm through the gateway using the private network (wireguard) as the backend.
+
+This will be done through the following steps:
+
+### Get the Client
+
+```ts
+const grid3 = getClient();
+```
+
+### Get the Nodes
+
+Determine the deploying nodes for the vm, network and gateway.
+
+- Gateway and network access node
+
+ ```ts
+ const gwNode = +(await grid3.capacity.filterNodes({ gateway: true }))[0].nodeId;
+ ```
+
+ Using the `filterNodes` method, will get the first gateway node id, we will deploy the gateway and will use it as our network access node.
+
+ > The gateway node must be the same as the network access node.
+- VM node
+
+ we need to set the filter options first for this example we will deploy the vm with 1 cpu, 2 GB of memory.
+ now will crete a `FilterOptions` object with that specs and get the firs node id of the result.
+
+ ```ts
+ const vmQueryOptions: FilterOptions = {
+ cru: 1,
+ mru: 2, // GB
+ availableFor: grid3.twinId,
+ farmId: 1,
+ };
+ const vmNode = +(await grid3.capacity.filterNodes(vmQueryOptions))[0].nodeId;
+ ```
+
+### Deploy the VM
+
+We need to create the network and machine models, the deploy the VM
+
+```ts
+const network = createNetworkModel(gwNode, "monNetwork");
+const vm = createMachineModel(vmNode);
+const machines = createMachinesModel(vm, network);
+log(`Deploying vm on node: ${vmNode}, with network node: ${gwNode}`);
+
+// deploy the vm
+const vmResult = await grid3.machines.deploy(machines);
+log(vmResult);
+```
+
+- `CreateNetWorkModel` :
+ we are creating a network and set the node id to be `gwNode`, the name `monNetwork` and inside the function we set `addAccess: true` to add __wireguard__ access.
+
+- `createMachineModel` and `createMachinesModel` is similar to the previous section of [deploying a single VM](../javascript/grid3_javascript_vm.md), but we are passing the created `NetworkModel` to the machines model and the entry point here runs a simple python server.
+
+### Deploy the Gateway
+
+Now we have our VM deployed with it's network, we need to make the gateway on the same node, same network and pointing to the VM's private IP address.
+
+- Get the VM's private IP address:
+
+ ```ts
+ const vmPrivateIP = (deployedVm as { interfaces: { ip: string }[] }[])[0].interfaces[0].ip;
+ ```
+
+- Create the Gateway name model:
+
+ ```ts
+ const gateway = createGwModel(gwNode, vmPrivateIP, network.name, "pyserver", 8000);
+ ```
+
+ This will create a `GatewayNameModel` with the following properties:
+
+ - `name` : the subdomain name
+ - `node_id` : the gateway node id
+ - `tls_passthrough: false`
+ - `backends: [`http://${ip}:${port}`]` : the private ip address and the port number of our machine
+ - `network: networkName` : the network name, we already created earlier.
+
+### Get the Deployments Information
+
+ ```ts
+ const deployedVm = await grid3.machines.getObj(machines.name);
+ log("+++ deployed vm +++");
+ log(deployedVm);
+
+ log("+++ Deployed gateway +++");
+ const deployedGw = await grid3.gateway.getObj(gateway.name);
+ log(deployedGw);
+ ```
+
+- `deployedVm` : is an array of one object contains the details about the vm deployment.
+
+ ```ts
+ [
+ {
+ version: 0,
+ contractId: 30658,
+ nodeId: 11,
+ name: 'testvm1',
+ created: 1686225126,
+ status: 'ok',
+ message: '',
+ flist: 'https://hub.grid.tf/tf-official-apps/threefoldtech-ubuntu-22.04.flist',
+ publicIP: null,
+ planetary: '302:9e63:7d43:b742:3582:a831:cd41:3f19',
+ interfaces: [ { network: 'monNetwork', ip: '10.238.2.2' } ],
+ capacity: { cpu: 1, memory: 2048 },
+ mounts: [],
+ env: {
+ SSH_KEY: 'ssh'
+ },
+ entrypoint: '/usr/bin/python3 -m http.server --bind ::',
+ metadata: '{"type":"vm","name":"newVMs","projectName":""}',
+ description: 'test deploying VMs with wireguard via ts grid3 client',
+ rootfs_size: 0,
+ corex: false
+ }
+ ]
+ ```
+
+- `deployedGw` : is an array of one object contains the details of the gateway name.
+
+ ```ts
+ [
+ {
+ version: 0,
+ contractId: 30659,
+ name: 'pyserver1',
+ created: 1686225139,
+ status: 'ok',
+ message: '',
+ type: 'gateway-name-proxy',
+ domain: 'pyserver1.gent02.dev.grid.tf',
+ tls_passthrough: false,
+ backends: [ 'http://10.238.2.2:8000' ],
+ metadata: '{"type":"gateway","name":"pyserver1","projectName":""}',
+ description: ''
+ }
+ ]
+ ```
+
+ Now we can access the vm using the `domain` that returned in the object.
+
+### Disconnect the Client
+
+finally we need to disconnect the client using `await grid3.disconnect();`
+
+### Delete the Deployments
+
+If we want to delete the deployments we can just do this:
+
+```ts
+ const deletedMachines = await grid3.machines.delete({ name: machines.name});
+ log(deletedMachines);
+
+ const deletedGW = await grid3.gateway.delete_name({ name: gateway.name});
+ log(deletedGW);
+```
+
+
+
+## Conclusion
+
+This section presented a detailed description on how to create a virtual machine with private IP using Wireguard and use it as a backend for a name gateway.
+
+If you have any questions, you can ask the ThreeFold community for help on the [ThreeFold Forum](http://forum.threefold.io/) or on the [ThreeFold Grid Tester Community](https://t.me/threefoldtesting) on Telegram.
\ No newline at end of file
diff --git a/collections/documentation/developers/javascript/sidebar.md b/collections/documentation/developers/javascript/sidebar.md
new file mode 100644
index 0000000..421992b
--- /dev/null
+++ b/collections/documentation/developers/javascript/sidebar.md
@@ -0,0 +1,11 @@
+- [Installation](@grid3_javascript_installation)
+- [Loading client](@grid3_javascript_loadclient)
+- [Deploy a VM](@grid3_javascript_vm)
+- [Capacity planning](@grid3_javascript_capacity_planning)
+- [Deploy multiple VMs](@grid3_javascript_vms)
+- [Deploy CapRover](@grid3_javascript_caprover)
+- [Gateways](@grid3_javascript_vm_gateways)
+- [Deploy a Kubernetes cluster](@grid3_javascript_kubernetes)
+- [Deploy a ZDB](@grid3_javascript_zdb)
+- [QSFS](@grid3_javascript_qsfs)
+- [Key Value Store](@grid3_javascript_kvstore)
diff --git a/collections/documentation/developers/proxy/commands.md b/collections/documentation/developers/proxy/commands.md
new file mode 100644
index 0000000..94e78ad
--- /dev/null
+++ b/collections/documentation/developers/proxy/commands.md
@@ -0,0 +1,127 @@
+Commands
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Work on Docs](#work-on-docs)
+- [To start the GridProxy server](#to-start-the-gridproxy-server)
+- [Run tests](#run-tests)
+
+***
+
+## Introduction
+
+The Makefile makes it easier to do mostly all the frequently commands needed to work on the project.
+
+## Work on Docs
+
+we are using [swaggo/swag](https://github.com/swaggo/swag) to generate swagger docs based on the annotation inside the code.
+
+- install swag executable binary
+
+ ```bash
+ go install github.com/swaggo/swag/cmd/swag@latest
+ ```
+
+- now if you check the binary directory inside go directory you will find the executable file.
+
+ ```bash
+ ls $(go env GOPATH)/bin
+ ```
+
+- to run swag you can either use the full path `$(go env GOPATH)/bin/swag` or export go binary to `$PATH`
+
+ ```bash
+ export PATH=$PATH:$(go env GOPATH)/bin
+ ```
+
+- use swag to format code comments.
+
+ ```bash
+ swag fmt
+ ```
+
+- update the docs
+
+ ```bash
+ swag init
+ ```
+
+- to parse external types from vendor
+
+ ```bash
+ swag init --parseVendor
+ ```
+
+- for a full generate docs command
+
+ ```bash
+ make docs
+ ```
+
+## To start the GridProxy server
+
+After preparing the postgres database you can `go run` the main file in `cmds/proxy_server/main.go` which responsible for starting all the needed server/clients.
+
+The server options
+
+| Option | Description |
+| ------------------ | ----------------------------------------------------------------------------------------------------------------------- |
+| -address | Server ip address (default `":443"`) |
+| -ca | certificate authority used to generate certificate (default `"https://acme-staging-v02.api.letsencrypt.org/directory"`) |
+| -cert-cache-dir | path to store generated certs in (default `"/tmp/certs"`) |
+| -domain | domain on which the server will be served |
+| -email | email address to generate certificate with |
+| -log-level | log level `[debug\|info\|warn\|error\|fatal\|panic]` (default `"info"`) |
+| -no-cert | start the server without certificate |
+| -postgres-db | postgres database |
+| -postgres-host | postgres host |
+| -postgres-password | postgres password |
+| -postgres-port | postgres port (default 5432) |
+| -postgres-user | postgres username |
+| -tfchain-url | tF chain url (default `"wss://tfchain.dev.grid.tf/ws"`) |
+| -relay-url | RMB relay url (default`"wss://relay.dev.grid.tf"`) |
+| -mnemonics | Dummy user mnemonics for relay calls |
+| -v | shows the package version |
+
+For a full server setup:
+
+```bash
+make restart
+```
+
+## Run tests
+
+There is two types of tests in the project
+
+- Unit Tests
+ - Found in `pkg/client/*_test.go`
+ - Run with `go test -v ./pkg/client`
+- Integration Tests
+ - Found in `tests/queries/`
+ - Run with:
+
+ ```bash
+ go test -v \
+ --seed 13 \
+ --postgres-host \
+ --postgres-db tfgrid-graphql \
+ --postgres-password postgres \
+ --postgres-user postgres \
+ --endpoint \
+ --mnemonics
+ ```
+
+ - Or to run a specific test you can append the previous command with
+
+ ```bash
+ -run
+ ```
+
+ You can found the TestName in the `tests/queries/*_test.go` files.
+
+To run all the tests use
+
+```bash
+make test-all
+```
diff --git a/collections/documentation/developers/proxy/contributions.md b/collections/documentation/developers/proxy/contributions.md
new file mode 100644
index 0000000..3960676
--- /dev/null
+++ b/collections/documentation/developers/proxy/contributions.md
@@ -0,0 +1,55 @@
+Contributions Guide
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Project structure](#project-structure)
+ - [Internal](#internal)
+ - [Pkg](#pkg)
+- [Writing tests](#writing-tests)
+
+***
+
+## Introduction
+
+We propose a quick guide to learn how to contribute.
+
+## Project structure
+
+The main structure of the code base is as follows:
+
+- `charts`: helm chart
+- `cmds`: includes the project Golang entrypoints
+- `docs`: project documentation
+- `internal`: contains the explorer API logic and the cert manager implementation, this where most of the feature work will be done
+- `pkg`: contains client implementation and shared libs
+- `tests`: integration tests
+- `tools`: DB tools to prepare the Postgres DB for testing and development
+- `rootfs`: ZOS root endpoint that will be mounted in the docker image
+
+### Internal
+
+- `explorer`: contains the explorer server logic:
+ - `db`: the db connection and operations
+ - `mw`: defines the generic action mount that will be be used as http handler
+- `certmanager`: logic to ensure certificates are available and up to date
+
+`server.go` includes the logic for all the API operations.
+
+### Pkg
+
+- `client`: client implementation
+- `types`: defines all the API objects
+
+## Writing tests
+
+Adding a new endpoint should be accompanied with a corresponding test. Ideally every change or bug fix should include a test to ensure the new behavior/fix is working as intended.
+
+Since these are integration tests, you need to first make sure that your local db is already seeded with the ncessary data. See tools [doc](./db_testing.md) for more information about how to prepare your db.
+
+Testing tools offer two clients that are the basic of most tests:
+
+- `local`: this client connects to the local db
+- `proxy client`: this client connects to the running local instance
+
+You need to start an instance of the server before running the tests. Check [here](./commands.md) for how to start.
diff --git a/collections/documentation/developers/proxy/database.md b/collections/documentation/developers/proxy/database.md
new file mode 100644
index 0000000..58c327a
--- /dev/null
+++ b/collections/documentation/developers/proxy/database.md
@@ -0,0 +1,21 @@
+Database
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Max Open Connections](#max-open-connections)
+
+***
+
+## Introduction
+
+The grid proxy has access to a postgres database containing information about the tfgrid, specifically information about grid nodes, farms, twins, and contracts.\
+The database is filled/updated by this [indexer](https://github.com/threefoldtech/tfchain_graphql).
+The grid proxy mainly retrieves information from the db with a few modifications for efficient retrieval (e.g. adding indices, caching node gpus, etc..).
+
+## Max Open Connections
+
+The postgres database can handle 100 open connections concurrently (that is the default value set by postgres), this number can be increased, depending on the infrastructure, by modifying it in the postgres.conf file where the db is deployed, or by executing the following query `ALTER system SET max_connections=size-of-connection`, but this requires a db restart to take effect.\
+The explorer creates a connection pool to the postgres db, with a max open pool connections set to a specific number (currently 80).\
+It's important to distinguish between the database max connections, and the max pool open connections, because if the pool did not have any constraints, it would try to open as many connections as it wanted, without any notion of the maximum connections the database accepts. It's the database responsibility then to accept or deny the connection.\
+This is why the max number of open pool connections is set to 80: It's below the max connections the database could handle (100), and it gives room for other actors outside of the explorer to open connections with the database.\
diff --git a/collections/documentation/developers/proxy/db_testing.md b/collections/documentation/developers/proxy/db_testing.md
new file mode 100644
index 0000000..60bffed
--- /dev/null
+++ b/collections/documentation/developers/proxy/db_testing.md
@@ -0,0 +1,45 @@
+DB for testing
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Run postgresql container](#run-postgresql-container)
+- [Create the DB](#create-the-db)
+ - [Method 1: Generate a db with relevant schema using the db helper tool:](#method-1-generate-a-db-with-relevant-schema-using-the-db-helper-tool)
+ - [Method 2: Fill the DB from a Production db dump file, for example if you have `dump.sql` file, you can run:](#method-2-fill-the-db-from-a-production-db-dump-file-for-example-if-you-have-dumpsql-file-you-can-run)
+
+***
+
+## Introduction
+
+We show how to use a database for testing.
+
+## Run postgresql container
+
+ ```bash
+ docker run --rm --name postgres \
+ -e POSTGRES_USER=postgres \
+ -e POSTGRES_PASSWORD=postgres \
+ -e POSTGRES_DB=tfgrid-graphql \
+ -p 5432:5432 -d postgres
+ ```
+
+## Create the DB
+you can either Generate a db with relevant schema to test things locally quickly, or load a previously taken DB dump file:
+
+### Method 1: Generate a db with relevant schema using the db helper tool:
+
+ ```bash
+ cd tools/db/ && go run . \
+ --postgres-host 127.0.0.1 \
+ --postgres-db tfgrid-graphql \
+ --postgres-password postgres \
+ --postgres-user postgres \
+ --reset \
+ ```
+
+### Method 2: Fill the DB from a Production db dump file, for example if you have `dump.sql` file, you can run:
+
+ ```bash
+ psql -h 127.0.0.1 -U postgres -d tfgrid-graphql < dump.sql
+ ```
diff --git a/collections/documentation/developers/proxy/explorer.md b/collections/documentation/developers/proxy/explorer.md
new file mode 100644
index 0000000..2651cae
--- /dev/null
+++ b/collections/documentation/developers/proxy/explorer.md
@@ -0,0 +1,38 @@
+The Grid Explorer
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Explorer Overview](#explorer-overview)
+- [Explorer Endpoints](#explorer-endpoints)
+
+***
+
+## Introduction
+
+The Grid Explorer is a rest API used to index a various information from the TFChain.
+
+## Explorer Overview
+
+- Due to limitations on indexing information from the blockchain, Complex inter-tables queries and limitations can't be applied directly on the chain.
+- Here comes the TFGridDB, a shadow database contains all the data on the chain which is being updated each 2 hours.
+- Then the explorer can apply a raw SQL queries on the database with all limitations and filtration needed.
+- The used technology to extract the info from the blockchain is Subsquid check the [repo](https://github.com/threefoldtech/tfchain_graphql).
+
+## Explorer Endpoints
+
+| HTTP Verb | Endpoint | Description |
+| --------- | --------------------------- | ---------------------------------- |
+| GET | `/contracts` | Show all contracts on the chain |
+| GET | `/farms` | Show all farms on the chain |
+| GET | `/gateways` | Show all gateway nodes on the grid |
+| GET | `/gateways/:node_id` | Get a single gateway node details |
+| GET | `/gateways/:node_id/status` | Get a single node status |
+| GET | `/nodes` | Show all nodes on the grid |
+| GET | `/nodes/:node_id` | Get a single node details |
+| GET | `/nodes/:node_id/status` | Get a single node status |
+| GET | `/stats` | Show the grid statistics |
+| GET | `/twins` | Show all the twins on the chain |
+| GET | `/nodes/:node_id/statistics`| Get a single node ZOS statistics |
+
+For the available filters on each node. check `/swagger/index.html` endpoint on the running instance.
diff --git a/collections/documentation/developers/proxy/production.md b/collections/documentation/developers/proxy/production.md
new file mode 100644
index 0000000..fe4e108
--- /dev/null
+++ b/collections/documentation/developers/proxy/production.md
@@ -0,0 +1,117 @@
+Running Proxy in Production
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Production Run](#production-run)
+- [To upgrade the machine](#to-upgrade-the-machine)
+- [Dockerfile](#dockerfile)
+- [Update helm package](#update-helm-package)
+- [Install the chart using helm package](#install-the-chart-using-helm-package)
+
+***
+
+## Introduction
+
+We show how to run grid proxy in production.
+
+## Production Run
+
+- Download the latest binary [here](https://github.com/threefoldtech/tfgrid-sdk-go/tree/development/grid-client)
+- add the execution permission to the binary and move it to the bin directory
+
+ ```bash
+ chmod +x ./gridproxy-server
+ mv ./gridproxy-server /usr/local/bin/gridproxy-server
+ ```
+
+- Add a new systemd service
+
+```bash
+cat << EOF > /etc/systemd/system/gridproxy-server.service
+[Unit]
+Description=grid proxy server
+After=network.target
+
+[Service]
+ExecStart=gridproxy-server --domain gridproxy.dev.grid.tf --email omar.elawady.alternative@gmail.com -ca https://acme-v02.api.letsencrypt.org/directory --postgres-host 127.0.0.1 --postgres-db db --postgres-password password --postgres-user postgres --mnemonics
+Type=simple
+Restart=always
+User=root
+Group=root
+
+[Install]
+WantedBy=multi-user.target
+Alias=gridproxy.service
+EOF
+```
+
+- enable the service
+
+ ```bash
+ systemctl enable gridproxy.service
+ ```
+
+- start the service
+
+ ```bash
+ systemctl start gridproxy.service
+ ```
+
+- check the status
+
+ ```bash
+ systemctl status gridproxy.service
+ ```
+
+- The command options:
+ - domain: the host domain which will generate ssl certificate to.
+ - email: the mail used to run generate the ssl certificate.
+ - ca: certificate authority server url, e.g.
+ - let's encrypt staging: `https://acme-staging-v02.api.letsencrypt.org/directory`
+ - let's encrypt production: `https://acme-v02.api.letsencrypt.org/directory`
+ - postgres -\*: postgres connection info.
+
+## To upgrade the machine
+
+- just replace the binary with the new one and apply
+
+```bash
+systemctl restart gridproxy-server.service
+```
+
+- it you have changes in the `/etc/systemd/system/gridproxy-server.service` you have to run this command first
+
+```bash
+systemctl daemon-reload
+```
+
+## Dockerfile
+
+To build & run dockerfile
+
+```bash
+docker build -t threefoldtech/gridproxy .
+docker run --name gridproxy -e POSTGRES_HOST="127.0.0.1" -e POSTGRES_PORT="5432" -e POSTGRES_DB="db" -e POSTGRES_USER="postgres" -e POSTGRES_PASSWORD="password" -e MNEMONICS="" threefoldtech/gridproxy
+```
+
+## Update helm package
+
+- Do `helm lint charts/gridproxy`
+- Regenerate the packages `helm package -u charts/gridproxy`
+- Regenerate index.yaml `helm repo index --url https://threefoldtech.github.io/tfgridclient_proxy/ .`
+- Push your changes
+
+## Install the chart using helm package
+
+- Adding the repo to your helm
+
+ ```bash
+ helm repo add gridproxy https://threefoldtech.github.io/tfgridclient_proxy/
+ ```
+
+- install a chart
+
+ ```bash
+ helm install gridproxy/gridproxy
+ ```
diff --git a/collections/documentation/developers/proxy/proxy.md b/collections/documentation/developers/proxy/proxy.md
new file mode 100644
index 0000000..7dc936c
--- /dev/null
+++ b/collections/documentation/developers/proxy/proxy.md
@@ -0,0 +1,149 @@
+ Introducing Grid Proxy
+
+ Table of Content
+
+- [About](#about)
+- [How to Use the Project](#how-to-use-the-project)
+- [Used Technologies \& Prerequisites](#used-technologies--prerequisites)
+- [Start for Development](#start-for-development)
+- [Setup for Production](#setup-for-production)
+- [Get and Install the Binary](#get-and-install-the-binary)
+- [Add as a Systemd Service](#add-as-a-systemd-service)
+
+***
+
+
+
+## About
+
+The TFGrid client Proxy acts as an interface to access information about the grid. It supports features such as filtering, limitation, and pagination to query the various entities on the grid like nodes, contracts and farms. Additionally the proxy can contact the required twin ID to retrieve stats about the relevant objects and performing ZOS calls.
+
+The proxy is used as the backend of several threefold projects like:
+
+- [Dashboard](../../dashboard/dashboard.md)
+
+
+
+## How to Use the Project
+
+If you don't want to care about setting up your instance you can use one of the live instances. each works against a different TFChain network.
+
+- Dev network:
+ - Swagger:
+- Qa network:
+ - Swagger:
+- Test network:
+ - Swagger:
+- Main network:
+ - Swagger:
+
+Or follow the [development guide](#start-for-development) to run yours.
+By default, the instance runs against devnet. to configure that you will need to config this while running the server.
+
+> Note: You may face some differences between each instance and the others. that is normal because each network is in a different stage of development and works correctly with others parts of the Grid on the same network.
+
+
+## Used Technologies & Prerequisites
+
+1. **GoLang**: Mainly the two parts of the project written in `Go 1.17`, otherwise you can just download the compiled binaries from github [releases](https://github.com/threefoldtech/tfgrid-sdk-go/releases)
+2. **Postgresql**: Used to load the TFGrid DB
+3. **Docker**: Containerize the running services such as Postgres and Redis.
+4. **Mnemonics**: Secret seeds for adummy identity to use for the relay client.
+
+For more about the prerequisites and how to set up and configure them. follow the [Setup guide](./setup.md)
+
+
+
+## Start for Development
+
+To start the services for development or testing make sure first you have all the [Prerequisites](#used-technologies--prerequisites).
+
+- Clone this repo
+
+ ```bash
+ git clone https://github.com/threefoldtech/tfgrid-sdk-go.git
+ cd tfgrid-sdk-go/grid-proxy
+ ```
+
+- The `Makefile` has all that you need to deal with Db, Explorer, Tests, and Docs.
+
+ ```bash
+ make help # list all the available subcommands.
+ ```
+
+- For a quick test explorer server.
+
+ ```bash
+ make all-start e=
+ ```
+
+ Now you can access the server at `http://localhost:8080`
+- Run the tests
+
+ ```bash
+ make test-all
+ ```
+
+- Generate docs.
+
+ ```bash
+ make docs
+ ```
+
+To run in development environment see [here](./db_testing.md) how to generate test db or load a db dump then use:
+
+```sh
+go run cmds/proxy_server/main.go --address :8080 --log-level debug -no-cert --postgres-host 127.0.0.1 --postgres-db tfgrid-graphql --postgres-password postgres --postgres-user postgres --mnemonics
+```
+
+Then visit `http://localhost:8080/`
+
+For more illustrations about the commands needed to work on the project, see the section [Commands](./commands.md). For more info about the project structure and contributions guidelines check the section [Contributions](./contributions.md).
+
+
+
+## Setup for Production
+
+## Get and Install the Binary
+
+- You can either build the project:
+
+ ```bash
+ make build
+ chmod +x cmd/proxy_server/server \
+ && mv cmd/proxy_server/server /usr/local/bin/gridproxy-server
+ ```
+
+- Or download a release:
+ Check the [releases](https://github.com/threefoldtech/tfgrid-sdk-go/releases) page and edit the next command with the chosen version.
+
+ ```bash
+ wget https://github.com/threefoldtech/tfgrid-sdk-go/releases/download/v1.6.7-rc2/tfgridclient_proxy_1.6.7-rc2_linux_amd64.tar.gz \
+ && tar -xzf tfgridclient_proxy_1.6.7-rc2_linux_amd64.tar.gz \
+ && chmod +x server \
+ && mv server /usr/local/bin/gridproxy-server
+ ```
+
+## Add as a Systemd Service
+
+- Create the service file
+
+ ```bash
+ cat << EOF > /etc/systemd/system/gridproxy-server.service
+ [Unit]
+ Description=grid proxy server
+ After=network.target
+
+ [Service]
+ ExecStart=gridproxy-server --domain gridproxy.dev.grid.tf --email omar.elawady.alternative@gmail.com -ca https://acme-v02.api.letsencrypt.org/directory --substrate wss://tfchain.dev.grid.tf/ws --postgres-host 127.0.0.1 --postgres-db db --postgres-password password --postgres-user postgres --mnemonics
+ Type=simple
+ Restart=always
+ User=root
+ Group=root
+
+ [Install]
+ WantedBy=multi-user.target
+ Alias=gridproxy.service
+ EOF
+ ```
+
diff --git a/collections/documentation/developers/proxy/proxy_readme.md b/collections/documentation/developers/proxy/proxy_readme.md
new file mode 100644
index 0000000..aaf4266
--- /dev/null
+++ b/collections/documentation/developers/proxy/proxy_readme.md
@@ -0,0 +1,25 @@
+Grid Proxy
+
+Welcome to the *Grid Proxy* section of the TFGrid Manual!
+
+In this comprehensive guide, we delve into the intricacies of the ThreeFold Grid Proxy, a fundamental component that empowers the ThreeFold Grid ecosystem.
+
+This section is designed to provide users, administrators, and developers with a detailed understanding of the TFGrid Proxy, offering step-by-step instructions for its setup, essential commands, and insights into its various functionalities.
+
+The Grid Proxy plays a pivotal role in facilitating secure and efficient communication between nodes within the ThreeFold Grid, contributing to the decentralized and autonomous nature of the network.
+
+Whether you are a seasoned ThreeFold enthusiast or a newcomer exploring the decentralized web, this manual aims to be your go-to resource for navigating the ThreeFold Grid Proxy landscape.
+
+To assist you on your journey, we have organized the content into distinct chapters below, covering everything from initial setup procedures and database testing to practical commands, contributions, and insights into the ThreeFold Explorer and the Grid Proxy Database functionalities.
+
+Table of Contents
+
+- [Introducing Grid Proxy](./proxy.md)
+- [Setup](./setup.md)
+- [DB Testing](./db_testing.md)
+- [Commands](./commands.md)
+- [Contributions](./contributions.md)
+- [Explorer](./explorer.md)
+- [Database](./database.md)
+- [Production](./production.md)
+- [Release](./release.md)
\ No newline at end of file
diff --git a/collections/documentation/developers/proxy/release.md b/collections/documentation/developers/proxy/release.md
new file mode 100644
index 0000000..5f5fe84
--- /dev/null
+++ b/collections/documentation/developers/proxy/release.md
@@ -0,0 +1,32 @@
+Release Grid-Proxy
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Steps](#steps)
+- [Debugging](#debugging)
+
+***
+
+## Introduction
+
+We show the steps to release a new version of the Grid Proxy.
+
+## Steps
+
+To release a new version of the Grid-Proxy component, follow these steps:
+
+Update the `appVersion` field in the `charts/Chart.yaml` file. This field should reflect the new version number of the release.
+
+The release process includes generating and pushing a Docker image with the latest GitHub tag. This step is automated through the `gridproxy-release.yml` workflow.
+
+Trigger the `gridproxy-release.yml` workflow by pushing the desired tag to the repository. This will initiate the workflow, which will generate the Docker image based on the tag and push it to the appropriate registry.
+
+## Debugging
+In the event that the workflow does not run automatically after pushing the tag and making the release, you can manually execute it using the GitHub Actions interface. Follow these steps:
+
+Go to the [GitHub Actions page](https://github.com/threefoldtech/tfgrid-sdk-go/actions/workflows/gridproxy-release.yml) for the Grid-Proxy repository.
+
+Locate the workflow named gridproxy-release.yml.
+
+Trigger the workflow manually by selecting the "Run workflow" option.
\ No newline at end of file
diff --git a/collections/documentation/developers/proxy/setup.md b/collections/documentation/developers/proxy/setup.md
new file mode 100644
index 0000000..fa8d07f
--- /dev/null
+++ b/collections/documentation/developers/proxy/setup.md
@@ -0,0 +1,50 @@
+Setup
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Install Golang](#install-golang)
+- [Docker](#docker)
+- [Postgres](#postgres)
+- [Get Mnemonics](#get-mnemonics)
+
+***
+
+## Introduction
+
+We show how to set up grid proxy.
+
+## Install Golang
+
+To install Golang, you can follow the official [guide](https://go.dev/doc/install).
+
+## Docker
+
+Docker is useful for running the TFGridDb in container environment. Read this to [install Docker engine](../../system_administrators/computer_it_basics/docker_basics.md#install-docker-desktop-and-docker-engine).
+
+Note: it will be necessary to follow step #2 in the previous article to run docker without sudo. if you want to avoid that. edit the docker commands in the `Makefile` and add sudo.
+
+## Postgres
+
+If you have docker installed you can run postgres on a container with:
+
+```bash
+make db-start
+```
+
+Then you can either load a dump of the database if you have one:
+
+```bash
+make db-dump p=~/dump.sql
+```
+
+or easier you can fill the database tables with randomly generated data with the script `tools/db/generate.go` to do that run:
+
+```bash
+make db-fill
+```
+
+## Get Mnemonics
+
+1. Install [polkadot extension](https://github.com/polkadot-js/extension) on your browser.
+2. Create a new account from the extension. It is important to save the seeds.
diff --git a/collections/documentation/developers/tfchain/farming_policies.md b/collections/documentation/developers/tfchain/farming_policies.md
new file mode 100644
index 0000000..e997b8f
--- /dev/null
+++ b/collections/documentation/developers/tfchain/farming_policies.md
@@ -0,0 +1,94 @@
+ Farming Policies
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Farming Policy Fields](#farming-policy-fields)
+- [Limits on linked policy](#limits-on-linked-policy)
+- [Creating a Policy](#creating-a-policy)
+- [Linking a policy to a Farm](#linking-a-policy-to-a-farm)
+
+***
+
+## Introduction
+
+A farming policy defines how farming rewards are handed out for nodes. Every node has a farming policy attached. A farming policy is either linked to a farm, in which case new nodes are given the farming policy of the farm they are in once they register themselves. Alternatively a farming policy can be a "default". These are not attached to a farm, but instead they are used for nodes registered in farms which don't have a farming policy. Multiple defaults can exist at the same time, and the most fitting should be chosen.
+
+## Farming Policy Fields
+
+A farming policy has the following fields:
+
+- id (used to link policies)
+- name
+- Default. This indicates if the policy can be used by any new node (if the parent farm does not have a dedicated attached policy). Essentially, a `Default` policy serves as a base which can be overriden per farm by linking a non default policy to said farm.
+- Reward tft per CU, SU and NU, IPV4
+- Minimal uptime needed in integer format (example 995)
+- Policy end (After this block number the policy can not be linked to new farms any more)
+- If this policy is immutable or not. Immutable policies can never be changed again
+
+Additionally, we also use the following fields, though those are only useful for `Default` farming policies:
+
+- Node needs to be certified
+- Farm needs to be certified (with certification level, which will be changed to an enum).
+
+In case a farming policy is not attached to a farm, new nodes will pick the most appropriate farming policy from the default ones. To decide which one to pick, they should be considered in order with most restrictive first until one matches. That means:
+
+- First check for the policy with highest farming certification (in the current case gold) and certified nodes
+- Then check for a policy with highest farming certification (in the current case gold) and non certified nodes
+- Check for policy without farming certification but certified nodes
+- Last check for a policy without any kind of certification
+
+Important here is that certification of a node only happens after it comes live for the first time. As such, when a node gets certified, farming certification needs to be re evaluated, but only if the currently attached farming policy on the node is a `Default` policy (as specifically linked policies have priority over default ones). When evaluating again, we first consider if we are eligible for the farming policy linked to the farm, if any.
+
+## Limits on linked policy
+
+When a council member attaches a policy to a farm, limits can be set. These limits define how much a policy can be used for nodes, before it becomes unusable and gets removed. The limits currently are:
+
+- Farming Policy ID: the ID of the farming policy which we want to limit to a farm.
+- CU. Every time a node is added in the farm, it's CU is calculated and deducted from this amount. If the amount drops below 0, the maximum amount of CU that can be attached to this policy is reached.
+- SU. Every time a node is added in the farm, it's SU is calculated and deducted from this amount. If the amount drops below 0, the maximum amount of SU that can be attached to this policy is reached.
+- End date. After this date the policy is not effective anymore and can't be used. It is removed from the farm and a default policy is used.
+- Certification. If set, only certified nodes can get this policy. Non certified nodes get a default policy.
+
+Once a limit is reached, the farming policy is removed from the farm, so new nodes will get one of the default policies until a new policy is attached to the farm.
+
+## Creating a Policy
+
+A council member can create a Farming Policy (DAO) in the following way:
+
+1: Open [PolkadotJS](https://polkadot.js.org/apps/?rpc=wss%3A%2F%2Ftfchain.grid.tf#/extrinsics) apps on the corresponding network and go to `Extrinsics`
+2: Now select the account to propose from (should be an account that's a council member).
+3: Select as action `dao` -> `propose`
+5: Set a `threshold` (amount of farmers to vote)
+6: Select an actions `tfgridModule` -> `createFarmingPolicy` and fill in all the fields.
+7: Create a forum post with the details of the farming policy and fill in the link of that post in the `link` field
+8: Give it some good `description`.
+9: Duration is optional (by default it's 7 days). A proposal cannot be closed before the duration is "expired". If you wish to set a duration, the duration should be expressed in number of blocks from `now`. For example, 2 hours is equal to 1200 blocks (blocktime is 6 seconds) in this case, the duration should be filled in as `1200`.
+10: If all the fields are filled in, click `Propose`, now Farmers can vote. A proposal can be closed manually once there are enough votes AND the proposal is expired. To close go to extrinsics -> `dao` -> `close` -> fill in proposal hash and index (both can be found in chainstate).
+
+All (su, cu, nu, ipv4) values should be expressed in units USD. Minimal uptime should be expressed as integer that represents an percentage (example: `95`).
+
+Policy end is optional (0 or some block number in the future). This is used for expiration.
+
+For reference:
+
+![image](./img/create_policy.png)
+
+## Linking a policy to a Farm
+
+First identify the policy ID to link to a farm. You can check for farming policies in [chainstate](https://polkadot.js.org/apps/?rpc=wss%3A%2F%2Ftfchain.grid.tf#/chainstate) -> `tfgridModule` -> `farmingPolciesMap`, start with ID 1 and increment with 1 until you find the farming policy which was created when the proposal was expired and closed.
+
+1: Open [PolkadotJS](https://polkadot.js.org/apps/?rpc=wss%3A%2F%2Ftfchain.grid.tf#/extrinsics) apps on the corresponding network and go to `Extrinsics`
+2: Now select the account to propose from (should be an account that's a council member).
+3: Select as proposal `dao` -> `propose`
+4: Set a `threshold` (amount of farmers to vote)
+5: Select an actions `tfgridModule` -> `attachPolicyToFarm` and fill in all the fields (FarmID and Limits).
+6: Limits contains a `farming_policy_id` (Required) and cu, su, end, node count (which are all optional). It also contains `node_certification`, if this is set to true only certified nodes can have this policy.
+7: Create a forum post with the details of why we want to link that farm to that policy and fill in the link of that post in the `link` field
+8: Give it some good `description`.
+9: Duration is optional (by default it's 7 days). A proposal cannot be closed before the duration is "expired". If you wish to set a duration, the duration should be expressed in number of blocks from `now`. For example, 2 hours is equal to 1200 blocks (blocktime is 6 seconds) in this case, the duration should be filled in as `1200`.
+10: If all the fields are filled in, click `Propose`, now Farmers can vote. A proposal can be closed manually once there are enough votes AND the proposal is expired. To close go to extrinsics -> `dao` -> `close` -> fill in proposal hash and index (both can be found in chainstate).
+
+For reference:
+
+![image](./img/attach.png)
diff --git a/collections/documentation/developers/tfchain/img/TF.png b/collections/documentation/developers/tfchain/img/TF.png
new file mode 100644
index 0000000..528b5d9
Binary files /dev/null and b/collections/documentation/developers/tfchain/img/TF.png differ
diff --git a/collections/documentation/developers/tfchain/img/attach.png b/collections/documentation/developers/tfchain/img/attach.png
new file mode 100644
index 0000000..96e3c5f
Binary files /dev/null and b/collections/documentation/developers/tfchain/img/attach.png differ
diff --git a/collections/documentation/developers/tfchain/img/close_proposal.png b/collections/documentation/developers/tfchain/img/close_proposal.png
new file mode 100644
index 0000000..07e66a2
Binary files /dev/null and b/collections/documentation/developers/tfchain/img/close_proposal.png differ
diff --git a/collections/documentation/developers/tfchain/img/create_contract.png b/collections/documentation/developers/tfchain/img/create_contract.png
new file mode 100644
index 0000000..f082e80
Binary files /dev/null and b/collections/documentation/developers/tfchain/img/create_contract.png differ
diff --git a/collections/documentation/developers/tfchain/img/create_policy.png b/collections/documentation/developers/tfchain/img/create_policy.png
new file mode 100644
index 0000000..fa344e7
Binary files /dev/null and b/collections/documentation/developers/tfchain/img/create_policy.png differ
diff --git a/collections/documentation/developers/tfchain/img/create_provider.png b/collections/documentation/developers/tfchain/img/create_provider.png
new file mode 100644
index 0000000..e8668a2
Binary files /dev/null and b/collections/documentation/developers/tfchain/img/create_provider.png differ
diff --git a/collections/documentation/developers/tfchain/img/propose_approve.png b/collections/documentation/developers/tfchain/img/propose_approve.png
new file mode 100644
index 0000000..667f66f
Binary files /dev/null and b/collections/documentation/developers/tfchain/img/propose_approve.png differ
diff --git a/collections/documentation/developers/tfchain/img/proposed_approve.png b/collections/documentation/developers/tfchain/img/proposed_approve.png
new file mode 100644
index 0000000..5202c4c
Binary files /dev/null and b/collections/documentation/developers/tfchain/img/proposed_approve.png differ
diff --git a/collections/documentation/developers/tfchain/img/query_provider.png b/collections/documentation/developers/tfchain/img/query_provider.png
new file mode 100644
index 0000000..de66d4c
Binary files /dev/null and b/collections/documentation/developers/tfchain/img/query_provider.png differ
diff --git a/collections/documentation/developers/tfchain/img/service_contract_approve.png b/collections/documentation/developers/tfchain/img/service_contract_approve.png
new file mode 100644
index 0000000..1e0d034
Binary files /dev/null and b/collections/documentation/developers/tfchain/img/service_contract_approve.png differ
diff --git a/collections/documentation/developers/tfchain/img/service_contract_bill.png b/collections/documentation/developers/tfchain/img/service_contract_bill.png
new file mode 100644
index 0000000..55e84fe
Binary files /dev/null and b/collections/documentation/developers/tfchain/img/service_contract_bill.png differ
diff --git a/collections/documentation/developers/tfchain/img/service_contract_cancel.png b/collections/documentation/developers/tfchain/img/service_contract_cancel.png
new file mode 100644
index 0000000..7669510
Binary files /dev/null and b/collections/documentation/developers/tfchain/img/service_contract_cancel.png differ
diff --git a/collections/documentation/developers/tfchain/img/service_contract_create.png b/collections/documentation/developers/tfchain/img/service_contract_create.png
new file mode 100644
index 0000000..69ec62a
Binary files /dev/null and b/collections/documentation/developers/tfchain/img/service_contract_create.png differ
diff --git a/collections/documentation/developers/tfchain/img/service_contract_id.png b/collections/documentation/developers/tfchain/img/service_contract_id.png
new file mode 100644
index 0000000..e49c396
Binary files /dev/null and b/collections/documentation/developers/tfchain/img/service_contract_id.png differ
diff --git a/collections/documentation/developers/tfchain/img/service_contract_reject.png b/collections/documentation/developers/tfchain/img/service_contract_reject.png
new file mode 100644
index 0000000..6235530
Binary files /dev/null and b/collections/documentation/developers/tfchain/img/service_contract_reject.png differ
diff --git a/collections/documentation/developers/tfchain/img/service_contract_set_fees.png b/collections/documentation/developers/tfchain/img/service_contract_set_fees.png
new file mode 100644
index 0000000..6cfa91a
Binary files /dev/null and b/collections/documentation/developers/tfchain/img/service_contract_set_fees.png differ
diff --git a/collections/documentation/developers/tfchain/img/service_contract_set_metadata.png b/collections/documentation/developers/tfchain/img/service_contract_set_metadata.png
new file mode 100644
index 0000000..e472145
Binary files /dev/null and b/collections/documentation/developers/tfchain/img/service_contract_set_metadata.png differ
diff --git a/collections/documentation/developers/tfchain/img/service_contract_state.png b/collections/documentation/developers/tfchain/img/service_contract_state.png
new file mode 100644
index 0000000..e824552
Binary files /dev/null and b/collections/documentation/developers/tfchain/img/service_contract_state.png differ
diff --git a/collections/documentation/developers/tfchain/img/service_contract_twin_from_account.png b/collections/documentation/developers/tfchain/img/service_contract_twin_from_account.png
new file mode 100644
index 0000000..293bad2
Binary files /dev/null and b/collections/documentation/developers/tfchain/img/service_contract_twin_from_account.png differ
diff --git a/collections/documentation/developers/tfchain/img/vote_proposal.png b/collections/documentation/developers/tfchain/img/vote_proposal.png
new file mode 100644
index 0000000..16111a0
Binary files /dev/null and b/collections/documentation/developers/tfchain/img/vote_proposal.png differ
diff --git a/collections/documentation/developers/tfchain/introduction.md b/collections/documentation/developers/tfchain/introduction.md
new file mode 100644
index 0000000..a983b68
--- /dev/null
+++ b/collections/documentation/developers/tfchain/introduction.md
@@ -0,0 +1,57 @@
+ThreeFold Chain
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Deployed instances](#deployed-instances)
+- [Create a TFChain twin](#create-a-tfchain-twin)
+- [Get your twin ID](#get-your-twin-id)
+
+***
+
+## Introduction
+
+ThreeFold blockchain (aka TFChain) serves as a registry for Nodes, Farms, Digital Twins and Smart Contracts.
+It is the backbone of [ZOS](https://github.com/threefoldtech/zos) and other components.
+
+## Deployed instances
+
+- Development network (Devnet):
+
+ - Polkadot UI: [https://polkadot.js.org/apps/?rpc=wss%3A%2F%2F/tfchain.dev.grid.tf#/explorer](https://polkadot.js.org/apps/?rpc=wss%3A%2F%2F/tfchain.dev.grid.tf#/explorer)
+ - Websocket url: `wss://tfchain.dev.grid.tf`
+ - GraphQL UI: [https://graphql.dev.grid.tf/graphql](https://graphql.dev.grid.tf/graphql)
+
+- QA testing network (QAnet):
+
+ - Polkadot UI: [https://polkadot.js.org/apps/?rpc=wss%3A%2F%2F/tfchain.qa.grid.tf#/explorer](https://polkadot.js.org/apps/?rpc=wss%3A%2F%2F/tfchain.qa.grid.tf#/explorer)
+ - Websocket url: `wss://tfchain.qa.grid.tf`
+ - GraphQL UI: [https://graphql.qa.grid.tf/graphql](https://graphql.qa.grid.tf/graphql)
+
+- Test network (Testnet):
+
+ - Polkadot UI: [https://polkadot.js.org/apps/?rpc=wss%3A%2F%2F/tfchain.test.grid.tf#/explorer](https://polkadot.js.org/apps/?rpc=wss%3A%2F%2F/tfchain.test.grid.tf#/explorer)
+ - Websocket url: `wss://tfchain.test.grid.tf`
+ - GraphQL UI: [https://graphql.test.grid.tf/graphql](https://graphql.test.grid.tf/graphql)
+
+- Production network (Mainnet):
+
+ - Polkadot UI: [https://polkadot.js.org/apps/?rpc=wss%3A%2F%2F/tfchain.grid.tf#/explorer](https://polkadot.js.org/apps/?rpc=wss%3A%2F%2F/tfchain.grid.tf#/explorer)
+ - Websocket url: `wss://tfchain.grid.tf`
+ - GraphQL UI: [https://graphql.grid.tf/graphql](https://graphql.grid.tf/graphql)
+
+## Create a TFChain twin
+
+A twin is a unique identifier linked to a specific account on a given TFChain network.
+Actually there are 2 ways to create a twin:
+
+- With the [Dashboard](../../dashboard/wallet_connector.md)
+ - a twin is automatically generated while creating a TFChain account
+- With the TFConnect app
+ - a twin is automatically generated while creating a farm (in this case the twin will be created on mainnet)
+
+## Get your twin ID
+
+One can retrieve the twin ID associated to his account going to `Developer` -> `Chain state` -> `tfgridModule` -> `twinIdByAccountID()`.
+
+![service_contract_twin_from_account](img/service_contract_twin_from_account.png)
diff --git a/collections/documentation/developers/tfchain/tfchain.md b/collections/documentation/developers/tfchain/tfchain.md
new file mode 100644
index 0000000..a575535
--- /dev/null
+++ b/collections/documentation/developers/tfchain/tfchain.md
@@ -0,0 +1,95 @@
+ ThreeFold Chain
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Twins](#twins)
+- [Farms](#farms)
+- [Nodes](#nodes)
+- [Node Contract](#node-contract)
+- [Rent Contract](#rent-contract)
+- [Name Contract](#name-contract)
+- [Contract billing](#contract-billing)
+- [Contract locking](#contract-locking)
+- [Contract grace period](#contract-grace-period)
+- [DAO](#dao)
+- [Farming Policies](#farming-policies)
+- [Node Connection price](#node-connection-price)
+- [Node Certifiers](#node-certifiers)
+
+***
+
+## Introduction
+
+ThreeFold Chain (TFChain) is the base layer for everything that interacts with the grid. Nodes, farms, users are registered on the chain. It plays the central role in achieving decentralised consensus between a user and Node to deploy a certain workload. A contract can be created on the chain that is essentially an agreement between a node and user.
+
+## Twins
+
+A twin is the central Identity object that is used for every entity that lives on the grid. A twin optionally has an IPV6 planetary network address which can be used for communication between twins no matter of the location they are in. A twin is coupled to a private/public keypair on chain. This keypair can hold TFT on TF Chain.
+
+## Farms
+
+A farm must be created before a Node can be booted. Every farms needs to have an unique name and is linked to the Twin that creates the farm. Once a farm is created, a unique ID is generated. This ID can be used to provide to the boot image of a Node.
+
+## Nodes
+
+When a node is booted for the first time, it registers itself on the chain and a unique identity is generated for this Node.
+
+## Node Contract
+
+A node contract is a contract between a user and a Node to deploy a certain workload. The contract is specified as following:
+
+```
+{
+ "contract_id": auto generated,
+ "node_id": unique id of the node,
+ "deployment_data": some additional deployment data
+ "deployment_hash": hash of the deployment definition signed by the user
+ "public_ips": number of public ips to attach to the deployment contract
+}
+```
+
+We don't save the raw workload definition on the chain but only a hash of the definition. After the contract is created, the user must send the raw deployment to the specified node in the contract. He can find where to send this data by looking up the Node's twin and contacting that twin over the planetary network.
+
+## Rent Contract
+
+A rent contract is also a contract between a user and a Node, but instead of being able to reserve a part of the node's capacity, the full capacity is rented. Once a rent contract is created on a Node by a user, only this user can deploy node contracts on this specific node. A discount of 50% is given if a the user wishes to rent the full capacity of a node by creating a rent contract. All node contracts deployed on a node where a user has a rent contract are free of use expect for the public ip's which can be added on a node contract.
+
+## Name Contract
+
+A name contract is a contract that specifies a unique name to be used on the grid's webgateways. Once a name contract is created, this name can be used as and entrypoint for an application on the grid.
+
+## Contract billing
+
+Every contract is billed every 1 hour on the chain, the amount that is due is deducted from the user's wallet every 24 hours or when the user cancels his contract. The total amount acrued in those 24 hours gets send to following destinations:
+
+- 10% goes to the threefold foundation
+- 5% goes to staking pool wallet (to be implemented in a later phase)
+- 50% goes to certified sales channel
+- 35% TFT gets burned
+
+See [pricing](../../../knowledge_base/cloud/pricing/pricing.md) for more information on how the cost for a contract is calculated.
+
+## Contract locking
+
+To not overload the chain with transfer events and others we choose to lock the amount due for a contract every hour and after 24 hours unlock the amount and deduct it in one go. This lock is saved on a user's account, if the user has multiple contracts the locked amount will be stacked.
+
+## Contract grace period
+
+When the owner of a contract runs out funds on his wallet to pay for his deployment, the contract goes in to a Grace Period state. The deployment, whatever that might be, will be unaccessible during this period to the user. When the wallet is funded with TFT again, the contract goes back to a normal operating state. If the grace period runs out (by default 2 weeks) the user's deployment and data will be deleted from the node.
+
+## DAO
+
+See [DAO](../../dashboard/tfchain/tf_dao.md) for more information on the DAO on TF Chain.
+
+## Farming Policies
+
+See [farming_policies](farming_policies.md) for more information on the farming policies on TF Chain.
+
+## Node Connection price
+
+A connection price is set to every new Node that boots on the Grid. This connection price influences the amount of TFT farmed in a period. The connection price set on a node is permanent. The DAO can propose the increase / decrease of the connection price. At the time of writing the connection price is set to $ 0.08. When the DAO proposes a connection price and the vote is passed, new nodes will attach to the new connection price.
+
+## Node Certifiers
+
+Node certifiers are entities who are allowed to set a node's certification level to `Certified`. The DAO can propose to add / remove entities that can certify nodes. This is usefull for allowing approved resellers of Threefold nodes to mark nodes as Certified. A certified node farms 25% more tokens than `Diy` a node.
\ No newline at end of file
diff --git a/collections/documentation/developers/tfchain/tfchain_external_service_contract.md b/collections/documentation/developers/tfchain/tfchain_external_service_contract.md
new file mode 100644
index 0000000..992186a
--- /dev/null
+++ b/collections/documentation/developers/tfchain/tfchain_external_service_contract.md
@@ -0,0 +1,142 @@
+External Service Contract: How to set and execute
+Table of Contents
+
+- [Introduction](#introduction)
+- [Step 1: Create the contract and get its unique ID](#step-1-create-contract--get-unique-id)
+- [Step 2: Fill contract](#step-2-fill-contract)
+- [Step 3: Both parties approve contract](#step-3-both-parties-approve-contract)
+- [Step 4: Bill for the service](#step-4-bill-for-the-service)
+- [Step 5: Cancel the contract](#step-5-cancel-the-contract)
+
+***
+
+
+# Introduction
+
+It is now possible to create a generic contract between two TFChain users (without restriction of account type) for some external service and bill for it.
+
+The initial scenario is when two parties, a service provider and a consumer of the service, want to use TFChain to automatically handle the billing/payment process for an agreement (in TFT) they want to make for a service which is external from the grid.
+This is actually a more direct and generic feature if we compare to the initial rewarding model where a service provider (or solution provider) is receiving TFT from a rewards distribution process, linked to a node contract and based on a cloud capacity consumption, which follows specific billing rules.
+
+The initial requirements are:
+- Both service and consumer need to have their respective twin created on TFChain (if not, see [here](tfchain.md#create-a-tfchain-twin) how to do it)
+- Consumer account needs to be funded (lack of funds will simply result in the contract cancelation while billed)
+
+In the following steps we detail the sequence of extrinsics that need to be called in TFChain Polkadot portal for setting up and executing such contract.
+
+Make sure to use right [links](tfchain.md#deployed-instances) depending on the targeted network.
+
+
+# Step 1: Create contract / Get unique ID
+
+## Create service contract
+
+The contract creation can be initiated by both service or consumer.
+In TFChain Polkadot portal, the one who iniciates the contract should go to `Developer` -> `Extrinsics` -> `smartContractModule` -> `serviceContractCreate()`, using the account he pretends to use in the contract, and select the corresponding service and consumer accounts before submiting the transaction.
+
+![service_contract_create](img/service_contract_create.png)
+
+Once executed the service contract is `Created` between the two parties and a unique ID is generated.
+
+## Last service contract ID
+
+To get the last generated service contract ID go to `Developer` -> `Chain state` -> `smartContractModule` -> `serviceContractID()`.
+
+![service_contract_id](img/service_contract_id.png)
+
+## Parse service contract
+
+To get the corresponding contract details, go to `Developer` -> `Chain state` -> `smartContractModule` -> `serviceContracts()` and provide the contract ID.
+You should see the following details:
+
+![service_contract_state](img/service_contract_state.png)
+
+Check if the contract fields are correct, especially the twin ID of both service and consumer, to be sure you get the right contract ID, referenced as `serviceContractId`.
+
+## Wrong contract ID ?
+
+If twin IDs are wrong ([how to get my twin ID?](tfchain.md#get-your-twin-id)) on service contract fields it means the contract does not correspond to the last created contract.
+In this case parse the last contracts on stack by decreasing `serviceContractId` and try to identify the right one; or the contract was simply not created so you should repeat the creation process and evaluate the error log.
+
+
+# Step 2: Fill contract
+
+Once created, the service contract must be filled with its relative `per hour` fees:
+- `baseFee` is the constant "per hour" price (in TFT) for the service.
+- `variableFee` is the maximum "per hour" amount (in TFT) that can be billed extra.
+
+To provide these values (only service can set fees), go to `Developer` -> `Extrinsics` -> `smartContractModule` -> `serviceContractSetFees()` specifying `serviceContractId`.
+
+![service_contract_set_fees](img/service_contract_set_fees.png)
+
+Some metadata (the description of the service for example) must be filled in a similar way (`Developer` -> `Extrinsics` -> `smartContractModule` -> `serviceContractSetMetadata()`).
+In this case service or consumer can set metadata.
+
+![service_contract_set_metadata](img/service_contract_set_metadata.png)
+
+The agreement will be automatically considered `Ready` when both metadata and fees are set (`metadata` not empty and `baseFee` greater than zero).
+Note that whenever this condition is not reached both extrinsics can still be called to modify agreement.
+You can check the contract status at each step of flow by parsing it as shown [here](#parse-service-contract).
+
+
+# Step 3: Both parties approve contract
+
+Now having the agreement ready the contract can be submited for approval.
+To approve the agreement, go to `Developer` -> `Extrinsics` -> `smartContractModule` -> `serviceContractApprove()` specifying `serviceContractId`.
+
+![service_contract_approve](img/service_contract_approve.png)
+
+To reject the agreement, go to `Developer` -> `Extrinsics` -> `smartContractModule` -> `serviceContractReject()` specifying `serviceContractId`.
+
+![service_contract_reject](img/service_contract_reject.png)
+
+The contract needs to be explicitly `Approved` by both service and consumer to be ready for billing.
+Before reaching this state, if one of the parties decides to call the rejection extrinsic, it will instantly lead to the cancelation of the contract (and its permanent removal).
+
+
+# Step 4: Bill for the service
+
+Once the contract is accepted by both it can be billed.
+
+## Send bill to consumer
+
+Only the service can bill the consumer going to `Developer` -> `Extrinsics` -> `smartContractModule` -> `serviceContractBill()` specifying `serviceContractId` and billing informations such as `variableAmount` and some `metadata`.
+
+![service_contract_bill](img/service_contract_bill.png)
+
+## Billing frequency
+
+⚠️ Important: because a service should not charge the user if it doesn't work, it is required that bills be send in less than 1 hour intervals.
+Any bigger interval will result in a bounded 1 hour bill (in other words, extra time will not be billed).
+It is the service responsability to bill on right frequency!
+
+## Amount due calculation
+
+When the bill is received, the chain calculates the bill amount based on the agreement values as follows:
+
+~~~
+amount = baseFee * T / 3600 + variableAmount
+~~~
+
+where `T` is the elapsed time, in seconds and bounded by 3600 (see [above](#billing-frequency)), since last effective billing operation occured.
+
+## Protection against draining
+
+Note that if `variableAmount` is too high (i.e `variableAmount > variableFee * T / 3600`) the billing extrinsic will fail.
+The `variableFee` value in the contract is interpreted as being "per hour" and acts as a protection mechanism to avoid consumer draining.
+Indeed, as it is technically possible for the service to send a bill every second, there would be no gain for that (unless overloading the chain uselessly).
+So it is also the service responsability to set a suitable `variableAmount` according to the billing frequency!
+
+## Billing considerations
+
+Then, if all goes well and no error is dispatched after submitting the transaction, the consumer pays for the due amount calculated from the bill (see calculation detail [above](#amount-due-calculation)).
+In practice the amount is transferred from the consumer twin account to the service twin account.
+Be aware that if the consumer is out of funds the billing will fail AND the contract will automatically be canceled.
+
+
+# Step 5: Cancel the contract
+
+At every moment of the flow since the contract is created it can be canceled (and definitively removed).
+Only the service or the consumer can do it going to `Developer` -> `Extrinsics` -> `smartContractModule` -> `serviceContractCancel()` specifying `serviceContractId`.
+
+![service_contract_cancel](img/service_contract_cancel.png)
diff --git a/collections/documentation/developers/tfchain/tfchain_solution_provider.md b/collections/documentation/developers/tfchain/tfchain_solution_provider.md
new file mode 100644
index 0000000..c027d16
--- /dev/null
+++ b/collections/documentation/developers/tfchain/tfchain_solution_provider.md
@@ -0,0 +1,81 @@
+Solution Provider
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Changes to Contract Creation](#changes-to-contract-creation)
+- [Creating a Provider](#creating-a-provider)
+- [Council needs to approve a provider before it can be used](#council-needs-to-approve-a-provider-before-it-can-be-used)
+
+***
+
+## Introduction
+
+> Note: While the solution provider program is still active, the plan is to discontinue the program in the near future. We will update the manual as we get more information. We currently do not accept new solution providers.
+
+A "solution" is something running on the grid, created by a community member. This can be brought forward to the council, who can vote on it to recognize it as a solution. On contract creation, a recognized solution can be referenced, in which case part of the payment goes toward the address coupled to the solution. On chain a solution looks as follows:
+
+- Description (should be some text, limited in length. Limit should be rather low, if a longer one is desired a link can be inserted. 160 characters should be enough imo).
+- Up to 5 payout addresses, each with a payout percentage. This is the percentage of the payout received by the associated address. The amount is deducted from the payout to the treasury and specified as percentage of the total contract cost. As such, the sum of these percentages can never exceed 50%. If this value is not 50%, the remainder is payed to the treasure. Example: 10% payout percentage to addr 1, 5% payout to addr 2. This means 15% goes to the 2 listed addresses combined and 35% goes to the treasury (instead of usual 50). Rest remains as is. If the cost would be 10TFT, 1TFT goes to the address1, 0.5TFT goes to address 2, 3.5TFT goes to the treasury, instead of the default 5TFT to the treasury
+- A unique code. This code is used to link a solution to the contract (numeric ID).
+
+This means contracts need to carry an optional solution code. If the code is not specified (default), the 50% goes entirely to the treasury (as is always the case today).
+
+A solution can be created by calling the extrinsic `smartContractModule` -> `createSolutionProvider` with parameters:
+
+- description
+- link (to website)
+- list of providers
+
+Provider:
+
+- who (account id)
+- take (amount of take this account should get) specified as an integer of max 50. example: 25
+
+A forum post should be created with the details of the created solution provider, the dao can vote to approve this or not. If the solution provider get's approved, it can be referenced on contract creation.
+
+Note that a solution can be deleted. In this case, existing contracts should fall back to the default behavior (i.e. if code not found -> default).
+
+## Changes to Contract Creation
+
+When creating a contract, a `solution_provider_id` can be passed. An error will be returned if an invalid or non-approved solution provider id is passed.
+
+## Creating a Provider
+
+Creating a provider is as easy as going to the [polkadotJS UI](https://polkadot.js.org/apps/?rpc=wss%3A%2F%2Ftfchain.dev.grid.tf#/extrinsics) (Currently only on devnet)
+
+Select module `SmartContractModule` -> `createSolutionProvider(..)`
+
+Fill in all the details, you can specify up to 5 target accounts which can have a take of the TFT generated from being a provider. Up to a total maximum of 50%. `Take` should be specified as a integer, example (`25`).
+
+Once this object is created, a forum post should be created here:
+
+![create](./img/create_provider.png)
+
+## Council needs to approve a provider before it can be used
+
+First propose the solution to be approved:
+
+![propose_approve](./img/propose_approve.png)
+
+After submission it should like like this:
+
+![proposed_approved](./img/proposed_approve.png)
+
+Now another member of the council needs to vote:
+
+![vote](./img/vote_proposal.png)
+
+After enough votes are reached, it can be closed:
+
+![close](./img/close_proposal.png)
+
+If the close was executed without error the solution should be approved and ready to be used
+
+Query the solution: `chainstate` -> `SmartContractModule` -> `solutionProviders`
+
+![query](./img/query_provider.png)
+
+Now the solution provider can be referenced on contract creation:
+
+![create](./img/create_contract.png)
diff --git a/collections/documentation/developers/tfcmd/tfcmd.md b/collections/documentation/developers/tfcmd/tfcmd.md
new file mode 100644
index 0000000..daa502a
--- /dev/null
+++ b/collections/documentation/developers/tfcmd/tfcmd.md
@@ -0,0 +1,15 @@
+TFCMD
+
+TFCMD (`tfcmd`) is a command line interface to interact and develop on Threefold Grid using command line.
+
+Consult the [ThreeFoldTech TFCMD repository](https://github.com/threefoldtech/tfgrid-sdk-go/tree/development/grid-cli) for the latest updates. Make sure to read the [basics](../../system_administrators/getstarted/tfgrid3_getstarted.md).
+
+Table of Contents
+
+- [Getting Started](./tfcmd_basics.md)
+- [Deploy a VM](./tfcmd_vm.md)
+- [Deploy Kubernetes](./tfcmd_kubernetes.md)
+- [Deploy ZDB](./tfcmd_zdbs.md)
+- [Gateway FQDN](./tfcmd_gateway_fqdn.md)
+- [Gateway Name](./tfcmd_gateway_name.md)
+- [Contracts](./tfcmd_contracts.md)
\ No newline at end of file
diff --git a/collections/documentation/developers/tfcmd/tfcmd_basics.md b/collections/documentation/developers/tfcmd/tfcmd_basics.md
new file mode 100644
index 0000000..8816eea
--- /dev/null
+++ b/collections/documentation/developers/tfcmd/tfcmd_basics.md
@@ -0,0 +1,67 @@
+TFCMD Getting Started
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Installation](#installation)
+- [Login](#login)
+- [Commands](#commands)
+- [Using TFCMD](#using-tfcmd)
+
+***
+
+## Introduction
+
+This section covers the basics on how to set up and use TFCMD (`tfcmd`).
+
+TFCMD is available as binaries. Make sure to download the latest release and to stay up to date with new releases.
+
+## Installation
+
+An easy way to use TFCMD is to download and extract the TFCMD binaries to your path.
+
+- Download latest release from [releases](https://github.com/threefoldtech/tfgrid-sdk-go/releases)
+ - ```
+ wget
+ ```
+- Extract the binaries
+ - ```
+ tar -xvf
+ ```
+- Move `tfcmd` to any `$PATH` directory:
+ ```bash
+ mv tfcmd /usr/local/bin
+ ```
+
+## Login
+
+Before interacting with Threefold Grid with `tfcmd` you should login with your mnemonics and specify the grid network:
+
+```console
+$ tfcmd login
+Please enter your mnemonics:
+Please enter grid network (main,test):
+```
+
+This validates your mnemonics and store your mnemonics and network to your default configuration dir.
+Check [UserConfigDir()](https://pkg.go.dev/os#UserConfigDir) for your default configuration directory.
+
+## Commands
+
+You can run the command `tfcmd help` at any time to access the help section. This will also display the available commands.
+
+| Command | Description |
+| ---------- | ---------------------------------------------------------- |
+| cancel | Cancel resources on Threefold grid |
+| completion | Generate the autocompletion script for the specified shell |
+| deploy | Deploy resources to Threefold grid |
+| get | Get a deployed resource from Threefold grid |
+| help | Help about any command |
+| login | Login with mnemonics to a grid network |
+| version | Get latest build tag |
+
+## Using TFCMD
+
+Once you've logged in, you can use commands to deploy workloads on the TFGrid. Read the next sections for more information on different types of workloads available with TFCMD.
+
+
diff --git a/collections/documentation/developers/tfcmd/tfcmd_contracts.md b/collections/documentation/developers/tfcmd/tfcmd_contracts.md
new file mode 100644
index 0000000..bb14c5d
--- /dev/null
+++ b/collections/documentation/developers/tfcmd/tfcmd_contracts.md
@@ -0,0 +1,99 @@
+Contracts
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Get](#get)
+ - [Get Contracts](#get-contracts)
+ - [Get Contract](#get-contract)
+- [Cancel](#cancel)
+ - [Optional Flags](#optional-flags)
+
+***
+
+## Introduction
+
+We explain how to handle contracts on the TFGrid with `tfcmd`.
+
+## Get
+
+### Get Contracts
+
+Get all contracts
+
+```bash
+tfcmd get contracts
+```
+
+Example:
+
+```console
+$ tfcmd get contracts
+5:13PM INF starting peer session=tf-1184566 twin=81
+Node contracts:
+ID Node ID Type Name Project Name
+50977 21 network vm1network vm1
+50978 21 vm vm1 vm1
+50980 14 Gateway Name gatewaytest gatewaytest
+
+Name contracts:
+ID Name
+50979 gatewaytest
+```
+
+### Get Contract
+
+Get specific contract
+
+```bash
+tfcmd get contract
+```
+
+Example:
+
+```console
+$ tfcmd get contract 50977
+5:14PM INF starting peer session=tf-1185180 twin=81
+5:14PM INF contract:
+{
+ "contract_id": 50977,
+ "twin_id": 81,
+ "state": "Created",
+ "created_at": 1702480020,
+ "type": "node",
+ "details": {
+ "nodeId": 21,
+ "deployment_data": "{\"type\":\"network\",\"name\":\"vm1network\",\"projectName\":\"vm1\"}",
+ "deployment_hash": "21adc91ef6cdc915d5580b3f12732ac9",
+ "number_of_public_ips": 0
+ }
+}
+```
+
+## Cancel
+
+Cancel specified contracts or all contracts.
+
+```bash
+tfcmd cancel contracts ... [Flags]
+```
+
+Example:
+
+```console
+$ tfcmd cancel contracts 50856 50857
+5:17PM INF starting peer session=tf-1185964 twin=81
+5:17PM INF contracts canceled successfully
+```
+
+### Optional Flags
+
+- all: cancel all twin's contracts.
+
+Example:
+
+```console
+$ tfcmd cancel contracts --all
+5:17PM INF starting peer session=tf-1185964 twin=81
+5:17PM INF contracts canceled successfully
+```
\ No newline at end of file
diff --git a/collections/documentation/developers/tfcmd/tfcmd_gateway_fqdn.md b/collections/documentation/developers/tfcmd/tfcmd_gateway_fqdn.md
new file mode 100644
index 0000000..538438f
--- /dev/null
+++ b/collections/documentation/developers/tfcmd/tfcmd_gateway_fqdn.md
@@ -0,0 +1,87 @@
+Gateway FQDN
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Deploy](#deploy)
+ - [Required Flags](#required-flags)
+ - [Optional Flags](#optional-flags)
+- [Get](#get)
+- [Cancel](#cancel)
+
+***
+
+## Introduction
+
+We explain how to use gateway fully qualified domain names on the TFGrid using `tfcmd`.
+
+## Deploy
+
+```bash
+tfcmd deploy gateway fqdn [flags]
+```
+
+### Required Flags
+
+- name: name for the gateway deployment also used for canceling the deployment. must be unique.
+- node: node id to deploy gateway on.
+- backends: list of backends the gateway will forward requests to.
+- fqdn: FQDN pointing to the specified node.
+
+### Optional Flags
+
+-tls: add TLS passthrough option (default false).
+
+Example:
+
+```console
+$ tfcmd deploy gateway fqdn -n gatewaytest --node 14 --backends http://93.184.216.34:80 --fqdn example.com
+3:34PM INF deploying gateway fqdn
+3:34PM INF gateway fqdn deployed
+```
+
+## Get
+
+```bash
+tfcmd get gateway fqdn
+```
+
+gateway is the name used when deploying gateway-fqdn using tfcmd.
+
+Example:
+
+```console
+$ tfcmd get gateway fqdn gatewaytest
+2:05PM INF gateway fqdn:
+{
+ "NodeID": 14,
+ "Backends": [
+ "http://93.184.216.34:80"
+ ],
+ "FQDN": "awady.gridtesting.xyz",
+ "Name": "gatewaytest",
+ "TLSPassthrough": false,
+ "Description": "",
+ "NodeDeploymentID": {
+ "14": 19653
+ },
+ "SolutionType": "gatewaytest",
+ "ContractID": 19653
+}
+```
+
+## Cancel
+
+```bash
+tfcmd cancel
+```
+
+deployment-name is the name of the deployment specified in while deploying using tfcmd.
+
+Example:
+
+```console
+$ tfcmd cancel gatewaytest
+3:37PM INF canceling contracts for project gatewaytest
+3:37PM INF gatewaytest canceled
+```
\ No newline at end of file
diff --git a/collections/documentation/developers/tfcmd/tfcmd_gateway_name.md b/collections/documentation/developers/tfcmd/tfcmd_gateway_name.md
new file mode 100644
index 0000000..a4c8191
--- /dev/null
+++ b/collections/documentation/developers/tfcmd/tfcmd_gateway_name.md
@@ -0,0 +1,88 @@
+Gateway Name
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Deploy](#deploy)
+ - [Required Flags](#required-flags)
+ - [Optional Flags](#optional-flags)
+- [Get](#get)
+- [Cancel](#cancel)
+
+***
+
+## Introduction
+
+We explain how to use gateway names on the TFGrid using `tfcmd`.
+
+## Deploy
+
+```bash
+tfcmd deploy gateway name [flags]
+```
+
+### Required Flags
+
+- name: name for the gateway deployment also used for canceling the deployment. must be unique.
+- backends: list of backends the gateway will forward requests to.
+
+### Optional Flags
+
+- node: node id gateway should be deployed on.
+- farm: farm id gateway should be deployed on, if set choose available node from farm that fits vm specs (default 1). note: node and farm flags cannot be set both.
+-tls: add TLS passthrough option (default false).
+
+Example:
+
+```console
+$ tfcmd deploy gateway name -n gatewaytest --node 14 --backends http://93.184.216.34:80
+3:34PM INF deploying gateway name
+3:34PM INF fqdn: gatewaytest.gent01.dev.grid.tf
+```
+
+## Get
+
+```bash
+tfcmd get gateway name
+```
+
+gateway is the name used when deploying gateway-name using tfcmd.
+
+Example:
+
+```console
+$ tfcmd get gateway name gatewaytest
+1:56PM INF gateway name:
+{
+ "NodeID": 14,
+ "Name": "gatewaytest",
+ "Backends": [
+ "http://93.184.216.34:80"
+ ],
+ "TLSPassthrough": false,
+ "Description": "",
+ "SolutionType": "gatewaytest",
+ "NodeDeploymentID": {
+ "14": 19644
+ },
+ "FQDN": "gatewaytest.gent01.dev.grid.tf",
+ "NameContractID": 19643,
+ "ContractID": 19644
+}
+```
+
+## Cancel
+
+```bash
+tfcmd cancel
+```
+
+deployment-name is the name of the deployment specified in while deploying using tfcmd.
+
+Example:
+
+```console
+$ tfcmd cancel gatewaytest
+3:37PM INF canceling contracts for project gatewaytest
+3:37PM INF gatewaytest canceled
+```
\ No newline at end of file
diff --git a/collections/documentation/developers/tfcmd/tfcmd_kubernetes.md b/collections/documentation/developers/tfcmd/tfcmd_kubernetes.md
new file mode 100644
index 0000000..9a7c2b1
--- /dev/null
+++ b/collections/documentation/developers/tfcmd/tfcmd_kubernetes.md
@@ -0,0 +1,147 @@
+Kubernetes
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Deploy](#deploy)
+ - [Required Flags](#required-flags)
+ - [Optional Flags](#optional-flags)
+- [Get](#get)
+- [Cancel](#cancel)
+
+***
+
+## Introduction
+
+In this section, we explain how to deploy Kubernetes workloads on the TFGrid using `tfcmd`.
+
+## Deploy
+
+```bash
+tfcmd deploy kubernetes [flags]
+```
+
+### Required Flags
+
+- name: name for the master node deployment also used for canceling the cluster deployment. must be unique.
+- ssh: path to public ssh key to set in the master node.
+
+### Optional Flags
+
+- master-node: node id master should be deployed on.
+- master-farm: farm id master should be deployed on, if set choose available node from farm that fits master specs (default 1). note: master-node and master-farm flags cannot be set both.
+- workers-node: node id workers should be deployed on.
+- workers-farm: farm id workers should be deployed on, if set choose available node from farm that fits master specs (default 1). note: workers-node and workers-farm flags cannot be set both.
+- ipv4: assign public ipv4 for master node (default false).
+- ipv6: assign public ipv6 for master node (default false).
+- ygg: assign yggdrasil ip for master node (default true).
+- master-cpu: number of cpu units for master node (default 1).
+- master-memory: master node memory size in GB (default 1).
+- master-disk: master node disk size in GB (default 2).
+- workers-number: number of workers nodes (default 0).
+- workers-ipv4: assign public ipv4 for each worker node (default false)
+- workers-ipv6: assign public ipv6 for each worker node (default false)
+- workers-ygg: assign yggdrasil ip for each worker node (default true)
+- workers-cpu: number of cpu units for each worker node (default 1).
+- workers-memory: memory size for each worker node in GB (default 1).
+- workers-disk: disk size in GB for each worker node (default 2).
+
+Example:
+
+```console
+$ tfcmd deploy kubernetes -n kube --ssh ~/.ssh/id_rsa.pub --master-node 14 --workers-number 2 --workers-node 14
+4:21PM INF deploying network
+4:22PM INF deploying cluster
+4:22PM INF master yggdrasil ip: 300:e9c4:9048:57cf:504f:c86c:9014:d02d
+```
+
+## Get
+
+```bash
+tfcmd get kubernetes
+```
+
+kubernetes is the name used when deploying kubernetes cluster using tfcmd.
+
+Example:
+
+```console
+$ tfcmd get kubernetes examplevm
+3:14PM INF k8s cluster:
+{
+ "Master": {
+ "Name": "kube",
+ "Node": 14,
+ "DiskSize": 2,
+ "PublicIP": false,
+ "PublicIP6": false,
+ "Planetary": true,
+ "Flist": "https://hub.grid.tf/tf-official-apps/threefoldtech-k3s-latest.flist",
+ "FlistChecksum": "c87cf57e1067d21a3e74332a64ef9723",
+ "ComputedIP": "",
+ "ComputedIP6": "",
+ "YggIP": "300:e9c4:9048:57cf:e8a0:662b:4e66:8faa",
+ "IP": "10.20.2.2",
+ "CPU": 1,
+ "Memory": 1024
+ },
+ "Workers": [
+ {
+ "Name": "worker1",
+ "Node": 14,
+ "DiskSize": 2,
+ "PublicIP": false,
+ "PublicIP6": false,
+ "Planetary": true,
+ "Flist": "https://hub.grid.tf/tf-official-apps/threefoldtech-k3s-latest.flist",
+ "FlistChecksum": "c87cf57e1067d21a3e74332a64ef9723",
+ "ComputedIP": "",
+ "ComputedIP6": "",
+ "YggIP": "300:e9c4:9048:57cf:66d0:3ee4:294e:d134",
+ "IP": "10.20.2.2",
+ "CPU": 1,
+ "Memory": 1024
+ },
+ {
+ "Name": "worker0",
+ "Node": 14,
+ "DiskSize": 2,
+ "PublicIP": false,
+ "PublicIP6": false,
+ "Planetary": true,
+ "Flist": "https://hub.grid.tf/tf-official-apps/threefoldtech-k3s-latest.flist",
+ "FlistChecksum": "c87cf57e1067d21a3e74332a64ef9723",
+ "ComputedIP": "",
+ "ComputedIP6": "",
+ "YggIP": "300:e9c4:9048:57cf:1ae5:cc51:3ffc:81e",
+ "IP": "10.20.2.2",
+ "CPU": 1,
+ "Memory": 1024
+ }
+ ],
+ "Token": "",
+ "NetworkName": "",
+ "SolutionType": "kube",
+ "SSHKey": "",
+ "NodesIPRange": null,
+ "NodeDeploymentID": {
+ "14": 22743
+ }
+}
+```
+
+## Cancel
+
+```bash
+tfcmd cancel
+```
+
+deployment-name is the name of the deployment specified in while deploying using tfcmd.
+
+Example:
+
+```console
+$ tfcmd cancel kube
+3:37PM INF canceling contracts for project kube
+3:37PM INF kube canceled
+```
\ No newline at end of file
diff --git a/collections/documentation/developers/tfcmd/tfcmd_vm.md b/collections/documentation/developers/tfcmd/tfcmd_vm.md
new file mode 100644
index 0000000..21e1471
--- /dev/null
+++ b/collections/documentation/developers/tfcmd/tfcmd_vm.md
@@ -0,0 +1,171 @@
+
+Deploy a VM
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Deploy](#deploy)
+ - [Flags](#flags)
+ - [Required Flags](#required-flags)
+ - [Optional Flags](#optional-flags)
+ - [Examples](#examples)
+ - [Deploy a VM without GPU](#deploy-a-vm-without-gpu)
+ - [Deploy a VM with GPU](#deploy-a-vm-with-gpu)
+- [Get](#get)
+ - [Get Example](#get-example)
+- [Cancel](#cancel)
+ - [Cancel Example](#cancel-example)
+- [Questions and Feedback](#questions-and-feedback)
+
+***
+
+# Introduction
+
+In this section, we explore how to deploy a virtual machine (VM) on the ThreeFold Grid using `tfcmd`.
+
+# Deploy
+
+You can deploy a VM with `tfcmd` using the following template accompanied by required and optional flags:
+
+```bash
+tfcmd deploy vm [flags]
+```
+
+## Flags
+
+When you use `tfcmd`, there are two required flags (`name` and `ssh`), while the other remaining flags are optional. Using such optional flags can be used to deploy a VM with a GPU for example or to set an IPv6 address and much more.
+
+### Required Flags
+
+- **name**: name for the VM deployment also used for canceling the deployment. The name must be unique.
+- **ssh**: path to public ssh key to set in the VM.
+
+### Optional Flags
+
+- **node**: node ID the VM should be deployed on.
+- **farm**: farm ID the VM should be deployed on, if set choose available node from farm that fits vm specs (default `1`). Note: node and farm flags cannot both be set.
+- **cpu**: number of cpu units (default `1`).
+- **disk**: size of disk in GB mounted on `/data`. If not set, no disk workload is made.
+- **entrypoint**: entrypoint for the VM FList (default `/sbin/zinit init`). Note: setting this without the flist option will fail.
+- **flist**: FList used in the VM (default `https://hub.grid.tf/tf-official-apps/threefoldtech-ubuntu-22.04.flist`). Note: setting this without the entrypoint option will fail.
+- **ipv4**: assign public ipv4 for the VM (default `false`).
+- **ipv6**: assign public ipv6 for the VM (default `false`).
+- **memory**: memory size in GB (default `1`).
+- **rootfs**: root filesystem size in GB (default `2`).
+- **ygg**: assign yggdrasil ip for the VM (default `true`).
+- **gpus**: assign a list of gpus' IDs to the VM. Note: setting this without the node option will fail.
+
+## Examples
+
+We present simple examples on how to deploy a virtual machine with or without a GPU using `tfcmd`.
+
+### Deploy a VM without GPU
+
+```console
+$ tfcmd deploy vm --name examplevm --ssh ~/.ssh/id_rsa.pub --cpu 2 --memory 4 --disk 10
+12:06PM INF deploying network
+12:06PM INF deploying vm
+12:07PM INF vm yggdrasil ip: 300:e9c4:9048:57cf:7da2:ac99:99db:8821
+```
+### Deploy a VM with GPU
+
+```console
+$ tfcmd deploy vm --name examplevm --ssh ~/.ssh/id_rsa.pub --cpu 2 --memory 4 --disk 10 --gpus '0000:0e:00.0/1882/543f' --gpus '0000:0e:00.0/1887/593f' --node 12
+12:06PM INF deploying network
+12:06PM INF deploying vm
+12:07PM INF vm yggdrasil ip: 300:e9c4:9048:57cf:7da2:ac99:99db:8821
+```
+
+# Get
+
+To get the VM, use the following template:
+
+```bash
+tfcmd get vm
+```
+
+Make sure to replace `` with the name of the VM specified using `tfcmd`.
+
+## Get Example
+
+In the following example, the name of the deployment to get is `examplevm`.
+
+```console
+$ tfcmd get vm examplevm
+3:20PM INF vm:
+{
+ "Name": "examplevm",
+ "NodeID": 15,
+ "SolutionType": "examplevm",
+ "SolutionProvider": null,
+ "NetworkName": "examplevmnetwork",
+ "Disks": [
+ {
+ "Name": "examplevmdisk",
+ "SizeGB": 10,
+ "Description": ""
+ }
+ ],
+ "Zdbs": [],
+ "Vms": [
+ {
+ "Name": "examplevm",
+ "Flist": "https://hub.grid.tf/tf-official-apps/threefoldtech-ubuntu-22.04.flist",
+ "FlistChecksum": "",
+ "PublicIP": false,
+ "PublicIP6": false,
+ "Planetary": true,
+ "Corex": false,
+ "ComputedIP": "",
+ "ComputedIP6": "",
+ "YggIP": "301:ad3a:9c52:98d1:cd05:1595:9abb:e2f1",
+ "IP": "10.20.2.2",
+ "Description": "",
+ "CPU": 2,
+ "Memory": 4096,
+ "RootfsSize": 2048,
+ "Entrypoint": "/sbin/zinit init",
+ "Mounts": [
+ {
+ "DiskName": "examplevmdisk",
+ "MountPoint": "/data"
+ }
+ ],
+ "Zlogs": null,
+ "EnvVars": {
+ "SSH_KEY": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDcGrS1RT36rHAGLK3/4FMazGXjIYgWVnZ4bCvxxg8KosEEbs/DeUKT2T2LYV91jUq3yibTWwK0nc6O+K5kdShV4qsQlPmIbdur6x2zWHPeaGXqejbbACEJcQMCj8szSbG8aKwH8Nbi8BNytgzJ20Ysaaj2QpjObCZ4Ncp+89pFahzDEIJx2HjXe6njbp6eCduoA+IE2H9vgwbIDVMQz6y/TzjdQjgbMOJRTlP+CzfbDBb6Ux+ed8F184bMPwkFrpHs9MSfQVbqfIz8wuq/wjewcnb3wK9dmIot6CxV2f2xuOZHgNQmVGratK8TyBnOd5x4oZKLIh3qM9Bi7r81xCkXyxAZbWYu3gGdvo3h85zeCPGK8OEPdYWMmIAIiANE42xPmY9HslPz8PAYq6v0WwdkBlDWrG3DD3GX6qTt9lbSHEgpUP2UOnqGL4O1+g5Rm9x16HWefZWMjJsP6OV70PnMjo9MPnH+yrBkXISw4CGEEXryTvupfaO5sL01mn+UOyE= abdulrahman@AElawady-PC\n"
+ },
+ "NetworkName": "examplevmnetwork"
+ }
+ ],
+ "QSFS": [],
+ "NodeDeploymentID": {
+ "15": 22748
+ },
+ "ContractID": 22748
+}
+```
+
+# Cancel
+
+To cancel your VM deployment, use the following template:
+
+```bash
+tfcmd cancel
+```
+
+Make sure to replace `` with the name of the deployment specified using `tfcmd`.
+
+## Cancel Example
+
+In the following example, the name of the deployment to cancel is `examplevm`.
+
+```console
+$ tfcmd cancel examplevm
+3:37PM INF canceling contracts for project examplevm
+3:37PM INF examplevm canceled
+```
+
+# Questions and Feedback
+
+If you have any questions or feedback, you can ask the ThreeFold community for help on the [ThreeFold Forum](http://forum.threefold.io/) or on the [ThreeFold Grid Tester Community](https://t.me/threefoldtesting) on Telegram.
\ No newline at end of file
diff --git a/collections/documentation/developers/tfcmd/tfcmd_zdbs.md b/collections/documentation/developers/tfcmd/tfcmd_zdbs.md
new file mode 100644
index 0000000..b9c01d7
--- /dev/null
+++ b/collections/documentation/developers/tfcmd/tfcmd_zdbs.md
@@ -0,0 +1,125 @@
+ZDBs
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Deploy](#deploy)
+ - [Required Flags](#required-flags)
+ - [Optional Flags](#optional-flags)
+- [Get](#get)
+- [Cancel](#cancel)
+
+***
+
+## Introduction
+
+In this section, we explore how to use ZDBs related commands using `tfcmd` to interact with the TFGrid.
+
+## Deploy
+
+```bash
+tfcmd deploy zdb [flags]
+```
+
+### Required Flags
+
+- project_name: project name for the ZDBs deployment also used for canceling the deployment. must be unique.
+- size: HDD of zdb in GB.
+
+### Optional Flags
+
+- node: node id zdbs should be deployed on.
+- farm: farm id zdbs should be deployed on, if set choose available node from farm that fits zdbs deployment specs (default 1). note: node and farm flags cannot be set both.
+- count: count of zdbs to be deployed (default 1).
+- names: a slice of names for the number of ZDBs.
+- password: password for ZDBs deployed
+- description: description for your ZDBs, it's optional.
+- mode: the enumeration of the modes 0-db can operate in (default user).
+- public: if zdb gets a public ip6 (default false).
+
+Example:
+
+- Deploying ZDBs
+
+```console
+$ tfcmd deploy zdb --project_name examplezdb --size=10 --count=2 --password=password
+12:06PM INF deploying zdbs
+12:06PM INF zdb 'examplezdb0' is deployed
+12:06PM INF zdb 'examplezdb1' is deployed
+```
+
+## Get
+
+```bash
+tfcmd get zdb
+```
+
+`zdb-project-name` is the name of the deployment specified in while deploying using tfcmd.
+
+Example:
+
+```console
+$ tfcmd get zdb examplezdb
+3:20PM INF zdb:
+{
+ "Name": "examplezdb",
+ "NodeID": 11,
+ "SolutionType": "examplezdb",
+ "SolutionProvider": null,
+ "NetworkName": "",
+ "Disks": [],
+ "Zdbs": [
+ {
+ "name": "examplezdb1",
+ "password": "password",
+ "public": false,
+ "size": 10,
+ "description": "",
+ "mode": "user",
+ "ips": [
+ "2a10:b600:1:0:c4be:94ff:feb1:8b3f",
+ "302:9e63:7d43:b742:469d:3ec2:ab15:f75e"
+ ],
+ "port": 9900,
+ "namespace": "81-36155-examplezdb1"
+ },
+ {
+ "name": "examplezdb0",
+ "password": "password",
+ "public": false,
+ "size": 10,
+ "description": "",
+ "mode": "user",
+ "ips": [
+ "2a10:b600:1:0:c4be:94ff:feb1:8b3f",
+ "302:9e63:7d43:b742:469d:3ec2:ab15:f75e"
+ ],
+ "port": 9900,
+ "namespace": "81-36155-examplezdb0"
+ }
+ ],
+ "Vms": [],
+ "QSFS": [],
+ "NodeDeploymentID": {
+ "11": 36155
+ },
+ "ContractID": 36155,
+ "IPrange": ""
+}
+```
+
+## Cancel
+
+```bash
+tfcmd cancel
+```
+
+`zdb-project-name` is the name of the deployment specified in while deploying using tfcmd.
+
+Example:
+
+```console
+$ tfcmd cancel examplezdb
+3:37PM INF canceling contracts for project examplezdb
+3:37PM INF examplezdb canceled
+```
\ No newline at end of file
diff --git a/collections/documentation/developers/tfrobot/tfrobot.md b/collections/documentation/developers/tfrobot/tfrobot.md
new file mode 100644
index 0000000..c8b2d5f
--- /dev/null
+++ b/collections/documentation/developers/tfrobot/tfrobot.md
@@ -0,0 +1,13 @@
+TFROBOT
+
+TFROBOT (`tfrobot`) is a command line interface tool that offers simultaneous mass deployment of groups of VMs on the ThreeFold Grid, with support of multiple retries for failed deployments, and customizable configurations, where you can define node groups, VMs groups and other configurations through a YAML or a JSON file.
+
+Consult the [ThreeFoldTech TFROBOT repository](https://github.com/threefoldtech/tfgrid-sdk-go/tree/development/tfrobot) for the latest updates and read the [basics](../../system_administrators/getstarted/tfgrid3_getstarted.md) to get up to speed if needed.
+
+Table of Contents
+
+- [Installation](./tfrobot_installation.md)
+- [Configuration File](./tfrobot_config.md)
+- [Deployment](./tfrobot_deploy.md)
+- [Commands and Flags](./tfrobot_commands_flags.md)
+- [Supported Configurations](./tfrobot_configurations.md)
\ No newline at end of file
diff --git a/collections/documentation/developers/tfrobot/tfrobot_commands_flags.md b/collections/documentation/developers/tfrobot/tfrobot_commands_flags.md
new file mode 100644
index 0000000..f33c59d
--- /dev/null
+++ b/collections/documentation/developers/tfrobot/tfrobot_commands_flags.md
@@ -0,0 +1,57 @@
+ Commands and Flags
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Commands](#commands)
+- [Subcommands](#subcommands)
+- [Flags](#flags)
+
+***
+
+## Introduction
+
+We present the various commands, subcommands and flags available with TFROBOT.
+
+
+## Commands
+
+You can run the command `tfrobot help` at any time to access the help section. This will also display the available commands.
+
+| Command | Description |
+| ---------- | ---------------------------------------------------------- |
+| completion | Generate the autocompletion script for the specified shell |
+| help | Help about any command |
+| version | Get latest build tag |
+
+Use `tfrobot [command] --help` for more information about a command.
+
+## Subcommands
+
+You can use subcommands to deploy and cancel workloads on the TFGrid.
+
+- **deploy:** used to mass deploy groups of vms with specific configurations
+ ```bash
+ tfrobot deploy -c path/to/your/config.yaml
+ ```
+- **cancel:** used to cancel all vms deployed using specific configurations
+ ```bash
+ tfrobot cancel -c path/to/your/config.yaml
+ ```
+- **load:** used to load all vms deployed using specific configurations
+ ```bash
+ tfrobot load -c path/to/your/config.yaml
+ ```
+
+## Flags
+
+You can use different flags to configure your deployment.
+
+| Flag | Usage |
+| :---: | :---: |
+| -c | used to specify path to configuration file |
+| -o | used to specify path to output file to store the output info in |
+| -d | allow debug logs to appear in the output logs |
+| -h | help |
+
+> **Note:** Make sure to use every flag once. If the flag is repeated, it will ignore all values and take the last value of the flag.`
\ No newline at end of file
diff --git a/collections/documentation/developers/tfrobot/tfrobot_config.md b/collections/documentation/developers/tfrobot/tfrobot_config.md
new file mode 100644
index 0000000..55c2850
--- /dev/null
+++ b/collections/documentation/developers/tfrobot/tfrobot_config.md
@@ -0,0 +1,131 @@
+ Configuration File
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Examples](#examples)
+ - [YAML Example](#yaml-example)
+ - [JSON Example](#json-example)
+- [Create a Configuration File](#create-a-configuration-file)
+
+***
+
+## Introduction
+
+To use TFROBOT, the user needs to create a YAML or a JSON configuration file that will contain the mass deployment information, such as the groups information, number of VMs to deploy how, the compute, storage and network resources needed, as well as the user's credentials, such as the SSH public key, the network (main, test, dev, qa) and the TFChain mnemonics.
+
+## Examples
+
+We present here a configuration file example that deploys 3 nodes with 2 vcores, 16GB of RAM, 100GB of SSD, 50GB of HDD and an IPv4 address. The same deployment is shown with a YAML file and with a JSON file. Parsing is based on file extension, TFROBOT will use JSON format if the file has a JSON extension and YAML format otherwise.
+
+You can use this example for guidance, and make sure to replace placeholders and adapt the groups based on your actual project details. To the minimum, `ssh_key1` should be replaced by the user SSH public key and `example-mnemonic` should be replaced by the user mnemonics.
+
+Note that if no IPs are specified as true (IPv4 or IPv6), an Yggdrasil IP address will automatically be assigned to the VM, as at least one IP should be set to allow an SSH connection to the VM.
+
+### YAML Example
+
+```
+node_groups:
+ - name: group_a
+ nodes_count: 3
+ free_cpu: 2
+ free_mru: 16
+ free_ssd: 100
+ free_hdd: 50
+ dedicated: false
+ public_ip4: true
+ public_ip6: false
+ certified: false
+ region: europe
+vms:
+ - name: examplevm123
+ vms_count: 5
+ node_group: group_a
+ cpu: 1
+ mem: 0.25
+ public_ip4: true
+ public_ip6: false
+ ssd:
+ - size: 15
+ mount_point: /mnt/ssd
+ flist: https://hub.grid.tf/tf-official-apps/base:latest.flist
+ entry_point: /sbin/zinit init
+ root_size: 0
+ ssh_key: example1
+ env_vars:
+ user: user1
+ pwd: 1234
+ssh_keys:
+ example1: ssh_key1
+mnemonic: example-mnemonic
+network: dev
+max_retries: 5
+```
+
+### JSON Example
+
+```
+{
+ "node_groups": [
+ {
+ "name": "group_a",
+ "nodes_count": 3,
+ "free_cpu": 2,
+ "free_mru": 16,
+ "free_ssd": 100,
+ "free_hdd": 50,
+ "dedicated": false,
+ "public_ip4": true,
+ "public_ip6": false,
+ "certified": false,
+ "region": europe,
+ }
+ ],
+ "vms": [
+ {
+ "name": "examplevm123",
+ "vms_count": 5,
+ "node_group": "group_a",
+ "cpu": 1,
+ "mem": 0.25,
+ "public_ip4": true,
+ "public_ip6": false,
+ "ssd": [
+ {
+ "size": 15,
+ "mount_point": "/mnt/ssd"
+ }
+ ],
+ "flist": "https://hub.grid.tf/tf-official-apps/base:latest.flist",
+ "entry_point": "/sbin/zinit init",
+ "root_size": 0,
+ "ssh_key": "example1",
+ "env_vars": {
+ "user": "user1",
+ "pwd": "1234"
+ }
+ }
+ ],
+ "ssh_keys": {
+ "example1": "ssh_key1"
+ },
+ "mnemonic": "example-mnemonic",
+ "network": "dev",
+ "max_retries": 5
+}
+```
+
+## Create a Configuration File
+
+You can start with the example above and adjust for your specific deployment needs.
+
+- Create directory
+ ```
+ mkdir tfrobot_deployments && cd $_
+ ```
+- Create configuration file and adjust with the provided example above
+ ```
+ nano config.yaml
+ ```
+
+Once you've set your configuration file, all that's left is to deploy on the TFGrid. Read the next section for more information on how to deploy with TFROBOT.
\ No newline at end of file
diff --git a/collections/documentation/developers/tfrobot/tfrobot_configurations.md b/collections/documentation/developers/tfrobot/tfrobot_configurations.md
new file mode 100644
index 0000000..7ceb867
--- /dev/null
+++ b/collections/documentation/developers/tfrobot/tfrobot_configurations.md
@@ -0,0 +1,68 @@
+ Supported Configurations
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Config File](#config-file)
+- [Node Group](#node-group)
+- [Vms Groups](#vms-groups)
+- [Disk](#disk)
+
+***
+
+## Introduction
+
+When deploying with TFROBOT, you can set different configurations allowing for personalized deployments.
+
+## Config File
+
+| Field | Description| Supported Values|
+| :---: | :---: | :---: |
+| [node_group](#node-group) | description of all resources needed for each node_group | list of structs of type node_group |
+| [vms](#vms-groups) | description of resources needed for deploying groups of vms belong to node_group | list of structs of type vms |
+| ssh_keys | map of ssh keys with key=name and value=the actual ssh key | map of string to string |
+| mnemonic | mnemonic of the user | should be valid mnemonic |
+| network | valid network of ThreeFold Grid networks | main, test, qa, dev |
+| max_retries | times of retries of failed node groups | positive integer |
+
+## Node Group
+
+| Field | Description| Supported Values|
+| :---: | :---: | :---: |
+| name | name of node_group | node group name should be unique |
+| nodes_count | number of nodes in node group| nonzero positive integer |
+| free_cpu | number of cpu of node | nonzero positive integer max = 32 |
+| free_mru | free memory in the node in GB | min = 0.25, max = 256 |
+| free_ssd | free ssd storage in the node in GB | positive integer value |
+| free_hdd | free hdd storage in the node in GB | positive integer value |
+| dedicated | are nodes dedicated | `true` or `false` |
+| public_ip4 | should the nodes have free ip v4 | `true` or `false` |
+| public_ip6 | should the nodes have free ip v6 | `true` or `false` |
+| certified | should the nodes be certified(if false the nodes could be certified or DIY) | `true` or `false` |
+| region | region could be the name of the continents the nodes are located in | africa, americas, antarctic, antarctic ocean, asia, europe, oceania, polar |
+
+## Vms Groups
+
+| Field | Description| Supported Values|
+| :---: | :---: | :---: |
+| name | name of vm group | string value with no special characters |
+| vms_count | number of vms in vm group| nonzero positive integer |
+| node_group | name of node_group the vm belongs to | should be defined in node_groups |
+| cpu | number of cpu for vm | nonzero positive integer max = 32 |
+| mem | free memory in the vm in GB | min = 0.25, max 256 |
+| planetary | should the vm have yggdrasil ip | `true` or `false` |
+| public_ip4 | should the vm have free ip v4 | `true` or `false` |
+| public_ip6 | should the vm have free ip v6 | `true` or `false` |
+| flist | should be a link to valid flist | valid flist url with `.flist` or `.fl` extension |
+| entry_point | entry point of the flist | path to the entry point in the flist |
+| ssh_key | key of ssh key defined in the ssh_keys map | should be valid ssh_key defined in the ssh_keys map |
+| env_vars | map of env vars | map of type string to string |
+| ssd | list of disks | should be of type disk|
+| root_size | root size in GB | 0 for default root size, max 10TB |
+
+## Disk
+
+| Field | Description| Supported Values|
+| :---: | :---: | :---: |
+| size | disk size in GB| positive integer min = 15 |
+| mount_point | disk mount point | path to mountpoint |
diff --git a/collections/documentation/developers/tfrobot/tfrobot_deploy.md b/collections/documentation/developers/tfrobot/tfrobot_deploy.md
new file mode 100644
index 0000000..7e16d12
--- /dev/null
+++ b/collections/documentation/developers/tfrobot/tfrobot_deploy.md
@@ -0,0 +1,59 @@
+
+
+ Deployment
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Prerequisites](#prerequisites)
+- [Deploy Workloads](#deploy-workloads)
+- [Delete Workloads](#delete-workloads)
+- [Logs](#logs)
+- [Using TFCMD with TFROBOT](#using-tfcmd-with-tfrobot)
+ - [Get Contracts](#get-contracts)
+
+***
+
+## Introduction
+
+We present how to deploy workloads on the ThreeFold Grid using TFROBOT.
+
+## Prerequisites
+
+To deploy workloads on the TFGrid with TFROBOT, you first need to [install TFROBOT](./tfrobot_installation.md) on your machine and create a [configuration file](./tfrobot_config.md).
+
+## Deploy Workloads
+
+Once you've installed TFROBOT and created a configuration file, you can deploy on the TFGrid with the following command. Make sure to indicate the path to your configuration file.
+
+```bash
+tfrobot deploy -c ./config.yaml
+```
+
+## Delete Workloads
+
+To delete the contracts, you can use the following line. Make sure to indicate the path to your configuration file.
+
+```bash
+tfrobot cancel -c ./config.yaml
+```
+
+## Logs
+
+To ensure a complete log history, append `2>&1 | tee path/to/log/file` to the command being executed.
+
+```bash
+tfrobot deploy -c ./config.yaml 2>&1 | tee path/to/log/file
+```
+
+## Using TFCMD with TFROBOT
+
+### Get Contracts
+
+The TFCMD tool works well with TFROBOT, as it can be used to query the TFGrid, for example you can see the contracts created by TFROBOT by running the TFCMD command, taking into consideration that you are using the same mnemonics and are on the same network:
+
+```bash
+tfcmd get contracts
+```
+
+For more information on TFCMD, [read the documentation](../tfcmd/tfcmd.md).
\ No newline at end of file
diff --git a/collections/documentation/developers/tfrobot/tfrobot_installation.md b/collections/documentation/developers/tfrobot/tfrobot_installation.md
new file mode 100644
index 0000000..deec2b8
--- /dev/null
+++ b/collections/documentation/developers/tfrobot/tfrobot_installation.md
@@ -0,0 +1,36 @@
+Installation
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Installation](#installation)
+
+***
+
+## Introduction
+
+This section covers the basics on how to install TFROBOT (`tfrobot`).
+
+TFROBOT is available as binaries. Make sure to download the latest release and to stay up to date with new releases.
+
+## Installation
+
+To install TFROBOT, simply download and extract the TFROBOT binaries to your path.
+
+- Create a new directory for `tfgrid-sdk-go`
+ ```
+ mkdir tfgrid-sdk-go
+ cd tfgrid-sdk-go
+ ```
+- Download latest release from [releases](https://github.com/threefoldtech/tfgrid-sdk-go/releases)
+ - ```
+ wget https://github.com/threefoldtech/tfgrid-sdk-go/releases/download/v0.14.4/tfgrid-sdk-go_Linux_x86_64.tar.gz
+ ```
+- Extract the binaries
+ - ```
+ tar -xvf tfgrid-sdk-go_Linux_x86_64.tar.gz
+ ```
+- Move `tfrobot` to any `$PATH` directory:
+ ```bash
+ mv tfrobot /usr/local/bin
+ ```
\ No newline at end of file
diff --git a/collections/documentation/documentation.md b/collections/documentation/documentation.md
new file mode 100644
index 0000000..1724121
--- /dev/null
+++ b/collections/documentation/documentation.md
@@ -0,0 +1,14 @@
+ ThreeFold Documentation
+
+The section contains all the practical information for farmers, developers and system administrators of the ThreeFold Grid.
+
+For complementary information on ThreeFold, refer to the [ThreeFold Knowledge Base](../knowledge_base/knowledge_base.md).
+
+Table of Contents
+
+- [Dashboard](./dashboard/dashboard.md)
+- [Developers](./developers/developers.md)
+- [Farmers](./farmers/farmers.md)
+- [System Administrators](./system_administrators/system_administrators.md)
+- [ThreeFold Token](./threefold_token/threefold_token.md)
+- [FAQ](./faq/faq.md)
\ No newline at end of file
diff --git a/collections/documentation/faq/faq.md b/collections/documentation/faq/faq.md
new file mode 100644
index 0000000..511c34d
--- /dev/null
+++ b/collections/documentation/faq/faq.md
@@ -0,0 +1,2480 @@
+ThreeFold FAQ
+
+Table of Contents
+
+- [GENERAL FAQ](#general-faq)
+ - [Basic Facts](#basic-facts)
+ - [What is the the ThreeFold blockchain?](#what-is-the-the-threefold-blockchain)
+ - [What is the architecture of the ThreeFold Grid in simple terms?](#what-is-the-architecture-of-the-threefold-grid-in-simple-terms)
+ - [What is the difference between Internet capacity and connectivity? Does ThreeFold replace my Internet service provider (ISP)?](#what-is-the-difference-between-internet-capacity-and-connectivity-does-threefold-replace-my-internet-service-provider-isp)
+ - [What are the priorities of ThreeFold (the Three P of ThreeFold)? ThreeFold is a Planet first project, what does it mean?](#what-are-the-priorities-of-threefold-the-three-p-of-threefold-threefold-is-a-planet-first-project-what-does-it-mean)
+ - [I want to help build the new Internet. How can I become a ThreeFold certified 3node partner?](#i-want-to-help-build-the-new-internet-how-can-i-become-a-threefold-certified-3node-partner)
+ - [How can I create a twin on the TF Grid?](#how-can-i-create-a-twin-on-the-tf-grid)
+ - [ThreeFold Communication](#threefold-communication)
+ - [Is there a ThreeFold app for mobile?](#is-there-a-threefold-app-for-mobile)
+ - [I want to reach the ThreeFold community. What are ThreeFold social links?](#i-want-to-reach-the-threefold-community-what-are-threefold-social-links)
+ - [Could we reach out someone for publishing research work on ThreeFold?](#could-we-reach-out-someone-for-publishing-research-work-on-threefold)
+ - [Who can I write to for a proposal? Where can I send a proposal email for a new partnership opportunity with ThreeFold?](#who-can-i-write-to-for-a-proposal-where-can-i-send-a-proposal-email-for-a-new-partnership-opportunity-with-threefold)
+ - [How can I track and follow the progress and development of ThreeFold?](#how-can-i-track-and-follow-the-progress-and-development-of-threefold)
+ - [Why do some forum posts need to be approved?](#why-do-some-forum-posts-need-to-be-approved)
+ - [The Technology of ThreeFold](#the-technology-of-threefold)
+ - [What is a 3Node?](#what-is-a-3node)
+ - [What is the difference between a 3node and a ThreeFold farm?](#what-is-the-difference-between-a-3node-and-a-threefold-farm)
+ - [What is Zero-OS from ThreeFold?](#what-is-zero-os-from-threefold)
+ - [ThreeFold uses Quantum Safe Storage technology, what does it mean?](#threefold-uses-quantum-safe-storage-technology-what-does-it-mean)
+ - [Quantum Safe File System (QSFS) allows for part of the storage to go down and it can self repair, however it’s still attached to a single VM and a single point of failure. Can a QSFS instance be reattached to another VM to recover it?](#quantum-safe-file-system-qsfs-allows-for-part-of-the-storage-to-go-down-and-it-can-self-repair-however-its-still-attached-to-a-single-vm-and-a-single-point-of-failure-can-a-qsfs-instance-be-reattached-to-another-vm-to-recover-it)
+ - [Where does the ThreeFold Explorer take its data from?](#where-does-the-threefold-explorer-take-its-data-from)
+ - [How can I use the Gridproxy to query information on the TF Grid?](#how-can-i-use-the-gridproxy-to-query-information-on-the-tf-grid)
+ - [How can I see the stats of the ThreeFold Grid?](#how-can-i-see-the-stats-of-the-threefold-grid)
+ - [What is the difference between a seed phrase (mnemonics) and an HEX secret?](#what-is-the-difference-between-a-seed-phrase-mnemonics-and-an-hex-secret)
+ - [Buying and Transacting TFT](#buying-and-transacting-tft)
+ - [How long does it take when you use the BSC-Stellar Bridge?](#how-long-does-it-take-when-you-use-the-bsc-stellar-bridge)
+ - [On my website, users can donate TFT on the Stellar Chain. Is there a way for users on my website to easily track the total sum of TFT donated?](#on-my-website-users-can-donate-tft-on-the-stellar-chain-is-there-a-way-for-users-on-my-website-to-easily-track-the-total-sum-of-tft-donated)
+ - [TF Connect App, TF Dashboard, GraphQL, Grix Proxy and Polkadot Substrate](#tf-connect-app-tf-dashboard-graphql-grix-proxy-and-polkadot-substrate)
+ - [Is there a way to create or import another wallet in TF Connect App?](#is-there-a-way-to-create-or-import-another-wallet-in-tf-connect-app)
+ - [I created a farm on the TF Chain. On the TF Connect App Farmer Migration section, my farm is under Other v3 farms, is this normal?](#i-created-a-farm-on-the-tf-chain-on-the-tf-connect-app-farmer-migration-section-my-farm-is-under-other-v3-farms-is-this-normal)
+ - [I am trying to access my wallet in the ThreeFold Connect App. It worked fine before, but now I just get a white screen. What does it mean and what can I do?](#i-am-trying-to-access-my-wallet-in-the-threefold-connect-app-it-worked-fine-before-but-now-i-just-get-a-white-screen-what-does-it-mean-and-what-can-i-do)
+ - [When I open the ThreeFold Connect App, I get the error: Error in initialization in Flagsmith. How can I fix this issue?](#when-i-open-the-threefold-connect-app-i-get-the-error-error-in-initialization-in-flagsmith-how-can-i-fix-this-issue)
+ - [Apart form the ThreeFold Connect App Wallet, how can I check my TFT balance?](#apart-form-the-threefold-connect-app-wallet-how-can-i-check-my-tft-balance)
+ - [Is it possible to export the transaction history of a wallet to a CSV file?](#is-it-possible-to-export-the-transaction-history-of-a-wallet-to-a-csv-file)
+ - [How can I use GraphQl to find information on the ThreeFold Grid?](#how-can-i-use-graphql-to-find-information-on-the-threefold-grid)
+ - [What are the different links to ThreeFold's Graph QL depending on the network?](#what-are-the-different-links-to-threefolds-graph-ql-depending-on-the-network)
+ - [How can I find 3Nodes with IPv6 addresses?](#how-can-i-find-3nodes-with-ipv6-addresses)
+ - [How can I use GraphQL to see contracts on my 3Nodes?](#how-can-i-use-graphql-to-see-contracts-on-my-3nodes)
+ - [How can I use Grid Proxy to find information on the ThreeFold Grid and 3Nodes?](#how-can-i-use-grid-proxy-to-find-information-on-the-threefold-grid-and-3nodes)
+ - [Who is hosting GraphQL and Grid Proxy on the ThreeFold Grid?](#who-is-hosting-graphql-and-grid-proxy-on-the-threefold-grid)
+ - [What is the difference between uptime, status and power state?](#what-is-the-difference-between-uptime-status-and-power-state)
+ - [I do not remember the name (ThreeFold 3bot ID) associated with my seed phrase on the ThreeFold Connect app. Can I recover my TF Connect app account with only the seed phrase and not the name (3bot ID) associated with it?](#i-do-not-remember-the-name-threefold-3bot-id-associated-with-my-seed-phrase-on-the-threefold-connect-app-can-i-recover-my-tf-connect-app-account-with-only-the-seed-phrase-and-not-the-name-3bot-id-associated-with-it)
+- [USERS FAQ](#users-faq)
+ - [TF Grid Functionalities](#tf-grid-functionalities)
+ - [What are the type of storage available on TF Grid?](#what-are-the-type-of-storage-available-on-tf-grid)
+ - [Deployments on the ThreeFold Grid](#deployments-on-the-threefold-grid)
+ - [Does the ThreeFold Grid charge the total resources rented or it only charges the resources used during deployment?](#does-the-threefold-grid-charge-the-total-resources-rented-or-it-only-charges-the-resources-used-during-deployment)
+ - [Do I pay for Internet traffic while deploying workloads on IPv4, IPv6 or Planetary Network?](#do-i-pay-for-internet-traffic-while-deploying-workloads-on-ipv4-ipv6-or-planetary-network)
+ - [What is the monthly cost for an IPv4 or an IPv6 public address on the ThreeFold Grid?](#what-is-the-monthly-cost-for-an-ipv4-or-an-ipv6-public-address-on-the-threefold-grid)
+ - [What are the differences between a container, a micro virtual machine and a full virtual machine (VM)?](#what-are-the-differences-between-a-container-a-micro-virtual-machine-and-a-full-virtual-machine-vm)
+ - [What is a 3Node gateway? How can I configure a 3Node as a gateway node?](#what-is-a-3node-gateway-how-can-i-configure-a-3node-as-a-gateway-node)
+ - [When connecting remotely with SSH, I get the following error: "WARNING: REMOTE HOST IDENTIFICATION HAS CHANGED!...". What can I do to fix this?](#when-connecting-remotely-with-ssh-i-get-the-following-error-warning-remote-host-identification-has-changed-what-can-i-do-to-fix-this)
+ - [How can I remove one host from known\_hosts?](#how-can-i-remove-one-host-from-known_hosts)
+ - [How can I add ThreeFold peers in the Yggdrasil configuration file?](#how-can-i-add-threefold-peers-in-the-yggdrasil-configuration-file)
+ - [How can I see Yggdrasil/Planetary Network's peers?](#how-can-i-see-yggdrasilplanetary-networks-peers)
+ - [How can I ping an Yggdrasil IP or IPv6 address?](#how-can-i-ping-an-yggdrasil-ip-or-ipv6-address)
+ - [Is there a way to test if I am properly connected to the Yggdrasil network (Planetary Network)?](#is-there-a-way-to-test-if-i-am-properly-connected-to-the-yggdrasil-network-planetary-network)
+ - [How can I change the username of my SSH key?](#how-can-i-change-the-username-of-my-ssh-key)
+ - [What is ThreeFold's stance on sharded workload? Will ThreeFold embrace and move towards distributed data chunks or stay with the one deployment, one node model?](#what-is-threefolds-stance-on-sharded-workload-will-threefold-embrace-and-move-towards-distributed-data-chunks-or-stay-with-the-one-deployment-one-node-model)
+ - [Tutorials and Guides](#tutorials-and-guides)
+ - [What is the minimum amount of TFT to deploy a Presearch node? How can I get a TFT discount when I deploy a Presearch node?](#what-is-the-minimum-amount-of-tft-to-deploy-a-presearch-node-how-can-i-get-a-tft-discount-when-i-deploy-a-presearch-node)
+ - [Can I use the same seed phrase for my mainnet and testnest accounts? How can I transfer my TFT from mainnet to testnet or vice versa?](#can-i-use-the-same-seed-phrase-for-my-mainnet-and-testnest-accounts-how-can-i-transfer-my-tft-from-mainnet-to-testnet-or-vice-versa)
+ - [Do I need a full or micro virtual machine (VM) when I run QSFS, quantum safe file system, on the ThreeFold Grid?](#do-i-need-a-full-or-micro-virtual-machine-vm-when-i-run-qsfs-quantum-safe-file-system-on-the-threefold-grid)
+ - [Linux, Github, Containers and More](#linux-github-containers-and-more)
+ - [Where should I start to learn more about Linux?](#where-should-i-start-to-learn-more-about-linux)
+ - [How can I clone a single branch of a repository on Github?](#how-can-i-clone-a-single-branch-of-a-repository-on-github)
+ - [Grace Period (Status Paused)](#grace-period-status-paused)
+ - [The status of my deployment is paused, in grace period, how can I resume the deployment?](#the-status-of-my-deployment-is-paused-in-grace-period-how-can-i-resume-the-deployment)
+ - [Once I refund my TF wallet, how long does it take for the deployment to resume from grace period?](#once-i-refund-my-tf-wallet-how-long-does-it-take-for-the-deployment-to-resume-from-grace-period)
+ - [Can I SSH into my deployments when they are in grace period (i.e. when their status is paused)?](#can-i-ssh-into-my-deployments-when-they-are-in-grace-period-ie-when-their-status-is-paused)
+ - [How long is the grace period (i.e. when the deployment status is paused)?](#how-long-is-the-grace-period-ie-when-the-deployment-status-is-paused)
+ - [Terraform](#terraform)
+ - [Working with Terraform, I get the following error: failed to create contract: ContractIsNotUnique. Is there a fix to this issue?](#working-with-terraform-i-get-the-following-error-failed-to-create-contract-contractisnotunique-is-there-a-fix-to-this-issue)
+ - [I am working with Terraform. What do I have to write in the file env.tfvars?](#i-am-working-with-terraform-what-do-i-have-to-write-in-the-file-envtfvars)
+ - [I am working with Terraform and I am using the example in Terraform Provider Grid. How can I use the example main.tf file with environment variables? Why am I getting the message Error: account not found, when deploying with Terraform?](#i-am-working-with-terraform-and-i-am-using-the-example-in-terraform-provider-grid-how-can-i-use-the-example-maintf-file-with-environment-variables-why-am-i-getting-the-message-error-account-not-found-when-deploying-with-terraform)
+ - [Users Troubleshooting and Error Messages](#users-troubleshooting-and-error-messages)
+ - [When deploying a virtual machine (VM) on the ThreeFold Grid, I get the following message after trying a full system update and upgrade: GRUB failed to install to the following devices... Is there a fix to this issue?](#when-deploying-a-virtual-machine-vm-on-the-threefold-grid-i-get-the-following-message-after-trying-a-full-system-update-and-upgrade-grub-failed-to-install-to-the-following-devices-is-there-a-fix-to-this-issue)
+ - [While deploying on the TF Dashboard, I get the following error :"global workload with the same name exists: conflict". What can I do to fix this issue?](#while-deploying-on-the-tf-dashboard-i-get-the-following-error-global-workload-with-the-same-name-exists-conflict-what-can-i-do-to-fix-this-issue)
+ - [ThreeFold Connect App](#threefold-connect-app)
+ - [TF Connect App is now asking for a 4-digit password (PIN). I don't remember it as I usually use touch or face ID to unlock the app. What can I do?](#tf-connect-app-is-now-asking-for-a-4-digit-password-pin-i-dont-remember-it-as-i-usually-use-touch-or-face-id-to-unlock-the-app-what-can-i-do)
+ - [Is there a way to have more than one wallet in TF Connect App?](#is-there-a-way-to-have-more-than-one-wallet-in-tf-connect-app)
+ - [What is the difference between 10.x.y.z and 192.168.x.y addresses?](#what-is-the-difference-between-10xyz-and-192168xy-addresses)
+- [DEVELOPERS FAQ](#developers-faq)
+ - [General Information for Developer](#general-information-for-developer)
+ - [Does Zero-OS assign private IPv4 addresses to workloads?](#does-zero-os-assign-private-ipv4-addresses-to-workloads)
+ - [Why does each 3Node server have two IP addresses associated with it?](#why-does-each-3node-server-have-two-ip-addresses-associated-with-it)
+ - [Can Zero-OS assign public IPv4 or IPv6 addresses to workloads?](#can-zero-os-assign-public-ipv4-or-ipv6-addresses-to-workloads)
+ - [What does MAC mean when it comes to networking?](#what-does-mac-mean-when-it-comes-to-networking)
+ - [I am a developer looking for a way to automatically convert BSC tokens into TFT. Could you please share tips on how to swap regular tokens into TFT, on backend, without and browser extensions, via any platform API?](#i-am-a-developer-looking-for-a-way-to-automatically-convert-bsc-tokens-into-tft-could-you-please-share-tips-on-how-to-swap-regular-tokens-into-tft-on-backend-without-and-browser-extensions-via-any-platform-api)
+ - [Test Net](#test-net)
+ - [Can I get some free TFT to test on Test Net](#can-i-get-some-free-tft-to-test-on-test-net)
+- [FARMERS FAQ](#farmers-faq)
+ - [TFT Farming Basics](#tft-farming-basics)
+ - [My Titan is v2.1 and the ThreeFold Grid is v3., what is the distinction?](#my-titan-is-v21-and-the-threefold-grid-is-v3-what-is-the-distinction)
+ - [When will I receive the farming rewards for my 3Nodes?](#when-will-i-receive-the-farming-rewards-for-my-3nodes)
+ - [What is the TFT minting process? Is it fully automated?](#what-is-the-tft-minting-process-is-it-fully-automated)
+ - [What should I do if I did not receive my farming rewards this month?](#what-should-i-do-if-i-did-not-receive-my-farming-rewards-this-month)
+ - [What is the TFT entry price of my 3Node farming rewards?](#what-is-the-tft-entry-price-of-my-3node-farming-rewards)
+ - [What is the necessary uptime for a 3Node per period of one month?](#what-is-the-necessary-uptime-for-a-3node-per-period-of-one-month)
+ - [How can I check the uptime of my 3Nodes? Is there a tool to check the uptime of 3Node servers on the ThreeFold Grid?](#how-can-i-check-the-uptime-of-my-3nodes-is-there-a-tool-to-check-the-uptime-of-3node-servers-on-the-threefold-grid)
+ - [I set up a 3Node in the middle of the month, does it affect uptime requirements and rewards?](#i-set-up-a-3node-in-the-middle-of-the-month-does-it-affect-uptime-requirements-and-rewards)
+ - [What is the difference between a certified and a non-certified 3Node?](#what-is-the-difference-between-a-certified-and-a-non-certified-3node)
+ - [What are the different certifications available for 3Node servers and farms? What are the Gold and Silver certifications?](#what-are-the-different-certifications-available-for-3node-servers-and-farms-what-are-the-gold-and-silver-certifications)
+ - [What is the difference between V2 and V3 minting?](#what-is-the-difference-between-v2-and-v3-minting)
+ - [What is the TFT minting address on Stellar Chain?](#what-is-the-tft-minting-address-on-stellar-chain)
+ - [Can Titans and DIY 3Nodes share the same farm?](#can-titans-and-diy-3nodes-share-the-same-farm)
+ - [Do I need one farm for each 3Node?](#do-i-need-one-farm-for-each-3node)
+ - [Can a single farm be composed of many 3Nodes?](#can-a-single-farm-be-composed-of-many-3nodes)
+ - [Can a single 3Node be on more than one farm?](#can-a-single-3node-be-on-more-than-one-farm)
+ - [Do I need one reward address per 3Node?](#do-i-need-one-reward-address-per-3node)
+ - [How can I access the expert bootstrap mode for Zero-OS?](#how-can-i-access-the-expert-bootstrap-mode-for-zero-os)
+ - [When it comes to the Zero-OS bootstrap image, can I simply duplicate the first image I burnt when I build another 3Node?](#when-it-comes-to-the-zero-os-bootstrap-image-can-i-simply-duplicate-the-first-image-i-burnt-when-i-build-another-3node)
+ - [If a node is unused for certain time (e.g. many months offline), will it be erased by the Grid?](#if-a-node-is-unused-for-certain-time-eg-many-months-offline-will-it-be-erased-by-the-grid)
+ - [Can a farm be erased from TF Grid?](#can-a-farm-be-erased-from-tf-grid)
+ - [On the ThreeFold Connect App, it says I need to migrate my Titan farm from V2 to V3. What do I have to do? How long does this take?](#on-the-threefold-connect-app-it-says-i-need-to-migrate-my-titan-farm-from-v2-to-v3-what-do-i-have-to-do-how-long-does-this-take)
+ - [How can I migrate my DIY farm from V2 to V3?](#how-can-i-migrate-my-diy-farm-from-v2-to-v3)
+ - [What does the pricing policy ID of a farm represent?](#what-does-the-pricing-policy-id-of-a-farm-represent)
+ - [What is the difference between TiB and TB? Why doesn't the TF Explorer shows the same storage space as my disk?](#what-is-the-difference-between-tib-and-tb-why-doesnt-the-tf-explorer-shows-the-same-storage-space-as-my-disk)
+ - [Farming Rewards and Related Notions](#farming-rewards-and-related-notions)
+ - [What are the rewards of farming? Can I get more rewards when my 3Node is being utilized?](#what-are-the-rewards-of-farming-can-i-get-more-rewards-when-my-3node-is-being-utilized)
+ - [How can I know the potential farming rewards for Grid Utilization?](#how-can-i-know-the-potential-farming-rewards-for-grid-utilization)
+ - [What is the easiest way to farm ThreeFold tokens (TFT)?](#what-is-the-easiest-way-to-farm-threefold-tokens-tft)
+ - [When do I receive my rewards?](#when-do-i-receive-my-rewards)
+ - [Do farming rewards take into account the type of RAM, SSD, HDD and CPU of the 3Node server?](#do-farming-rewards-take-into-account-the-type-of-ram-ssd-hdd-and-cpu-of-the-3node-server)
+ - [Can I send my farming rewards directly to a crypto exchange?](#can-i-send-my-farming-rewards-directly-to-a-crypto-exchange)
+ - [Do I need collateral to farm ThreeFold tokens?](#do-i-need-collateral-to-farm-threefold-tokens)
+ - [Can I add external drives to the 3Nodes to increase rewards and resources available to the ThreeFold Grid?](#can-i-add-external-drives-to-the-3nodes-to-increase-rewards-and-resources-available-to-the-threefold-grid)
+ - [Do I have access to the TFT rewards I receive each month when farming?](#do-i-have-access-to-the-tft-rewards-i-receive-each-month-when-farming)
+ - [What is TFTA? Is it still used?](#what-is-tfta-is-it-still-used)
+ - [Is there a way to certify a DIY 3Node? How can I become a 3Node certified vendor and builder?](#is-there-a-way-to-certify-a-diy-3node-how-can-i-become-a-3node-certified-vendor-and-builder)
+ - [Does it make sense to make my farm a company?](#does-it-make-sense-to-make-my-farm-a-company)
+ - [What is the difference between uptime and downtime, and between online and offline, when it comes to 3Nodes?](#what-is-the-difference-between-uptime-and-downtime-and-between-online-and-offline-when-it-comes-to-3nodes)
+ - [My 3Node server grid utilization is low, is it normal?](#my-3node-server-grid-utilization-is-low-is-it-normal)
+ - [3Node Farming Requirements](#3node-farming-requirements)
+ - [Can I host more than one 3Node server at my house?](#can-i-host-more-than-one-3node-server-at-my-house)
+ - [Is Wifi supported? Can I farm via Wifi instead of an Ethernet cable?](#is-wifi-supported-can-i-farm-via-wifi-instead-of-an-ethernet-cable)
+ - [I have 2 routers with each a different Internet service provider. I disconnected the ethernet cable from one router and connected it to the other router. Do I need to reboot the 3Node?](#i-have-2-routers-with-each-a-different-internet-service-provider-i-disconnected-the-ethernet-cable-from-one-router-and-connected-it-to-the-other-router-do-i-need-to-reboot-the-3node)
+ - [Do I need any specific port configuration when booting a 3Node?](#do-i-need-any-specific-port-configuration-when-booting-a-3node)
+ - [How much electricity does a 3Node use?](#how-much-electricity-does-a-3node-use)
+ - [Has anyone run stress tests to know the power consumption at heavy load of certain 3Nodes?](#has-anyone-run-stress-tests-to-know-the-power-consumption-at-heavy-load-of-certain-3nodes)
+ - [Can the Titan 3Node be run on PoE? (Power Over Ethernet)](#can-the-titan-3node-be-run-on-poe-power-over-ethernet)
+ - [What is the relationship between the 3Node's resources and bandwidth?](#what-is-the-relationship-between-the-3nodes-resources-and-bandwidth)
+ - [What is the bandwidth needed when it comes to running 3Nodes on the Grid?](#what-is-the-bandwidth-needed-when-it-comes-to-running-3nodes-on-the-grid)
+ - [Can I run Zero-OS on a virtual machine?](#can-i-run-zero-os-on-a-virtual-machine)
+ - [Is it possible to build a DIY 3Node with VMWare VM ?](#is-it-possible-to-build-a-diy-3node-with-vmware-vm-)
+ - [Can I run a 3Node on another operating system, like Windows, MAC or Linux?](#can-i-run-a-3node-on-another-operating-system-like-windows-mac-or-linux)
+ - [What is the minimum SSD requirement for a 3Node server to farm ThreeFold tokens (TFT)?](#what-is-the-minimum-ssd-requirement-for-a-3node-server-to-farm-threefold-tokens-tft)
+ - [Is it possible to have a 3Node server running on only HDD disks?](#is-it-possible-to-have-a-3node-server-running-on-only-hdd-disks)
+ - [Building a 3Node - Steps and Details](#building-a-3node---steps-and-details)
+ - [How can I be sure that I properly wiped my disks?](#how-can-i-be-sure-that-i-properly-wiped-my-disks)
+ - [If I wipe my disk to create a new node ID, will I lose my farming rewards during the month?](#if-i-wipe-my-disk-to-create-a-new-node-id-will-i-lose-my-farming-rewards-during-the-month)
+ - [My disks have issues with Zero-OS and my 3Nodes. How can I do a factory reset of the disks?](#my-disks-have-issues-with-zero-os-and-my-3nodes-how-can-i-do-a-factory-reset-of-the-disks)
+ - [Before doing a bootstrap image, I need to format my USB key. How can I format my USB key?](#before-doing-a-bootstrap-image-i-need-to-format-my-usb-key-how-can-i-format-my-usb-key)
+ - [What do you use to burn (or to load) the Zero-OS bootstrap image onto a USB stick?](#what-do-you-use-to-burn-or-to-load-the-zero-os-bootstrap-image-onto-a-usb-stick)
+ - [Should I do a UEFI image or a BIOS image to bootstrap Zero-OS?](#should-i-do-a-uefi-image-or-a-bios-image-to-bootstrap-zero-os)
+ - [How do I set the BIOS or UEFI of my 3Node?](#how-do-i-set-the-bios-or-uefi-of-my-3node)
+ - [For my 3Node server, do I need to enable virtualization in BIOS or UEFI?](#for-my-3node-server-do-i-need-to-enable-virtualization-in-bios-or-uefi)
+ - [How can I boot a 3Node server with a Zero-OS bootstrap image?](#how-can-i-boot-a-3node-server-with-a-zero-os-bootstrap-image)
+ - [The first time I booted my 3Node server, it says that the node is not registered yet. What can I do?](#the-first-time-i-booted-my-3node-server-it-says-that-the-node-is-not-registered-yet-what-can-i-do)
+ - [The first time I boot my 3Node, the node gets registered but it says cache disk : no ssd. What can I do?](#the-first-time-i-boot-my-3node-the-node-gets-registered-but-it-says-cache-disk--no-ssd-what-can-i-do)
+ - [The first time I boot my 3 node, the node gets registered and it says cache disk : OK, but the table System Used Capacity is empty. What can I do?](#the-first-time-i-boot-my-3-node-the-node-gets-registered-and-it-says-cache-disk--ok-but-the-table-system-used-capacity-is-empty-what-can-i-do)
+ - [I have a relatively old server (e.g. Dell R710 or R620, Z840). I have trouble booting Zero-OS. What could I do?](#i-have-a-relatively-old-server-eg-dell-r710-or-r620-z840-i-have-trouble-booting-zero-os-what-could-i-do)
+ - [I connected a SATA SSD to a CD-DVD optical drive adaptor. My system does not recognize the disk. What can I do?](#i-connected-a-sata-ssd-to-a-cd-dvd-optical-drive-adaptor-my-system-does-not-recognize-the-disk-what-can-i-do)
+ - [Can someone explain what should I put in the Public IP part of my farm? Should I just insert my Public IP and Gateway (given by my ISP)?](#can-someone-explain-what-should-i-put-in-the-public-ip-part-of-my-farm-should-i-just-insert-my-public-ip-and-gateway-given-by-my-isp)
+ - [Farming Optimization](#farming-optimization)
+ - [What is the difference between a ThreeFold 3Node and a ThreeFold farm? What is the difference between the farm ID and the node ID?](#what-is-the-difference-between-a-threefold-3node-and-a-threefold-farm-what-is-the-difference-between-the-farm-id-and-the-node-id)
+ - [How can I know how many GB of SSD and RAM do I need?](#how-can-i-know-how-many-gb-of-ssd-and-ram-do-i-need)
+ - [What is the optimal ratio of virtual cores (vcores or threads), SSD storage and RAM memory? What is the best optimization scenario for a 3Node, in terms of ThreeFold tokens (TFT) farming rewards?](#what-is-the-optimal-ratio-of-virtual-cores-vcores-or-threads-ssd-storage-and-ram-memory-what-is-the-best-optimization-scenario-for-a-3node-in-terms-of-threefold-tokens-tft-farming-rewards)
+ - [What does TBW mean? What is a good TBW level for a SSD disk?](#what-does-tbw-mean-what-is-a-good-tbw-level-for-a-ssd-disk)
+ - [Are SATA and SAS drives interchangeable?](#are-sata-and-sas-drives-interchangeable)
+ - [What is the speed difference between SAS and SATA disks?](#what-is-the-speed-difference-between-sas-and-sata-disks)
+ - [Is it possible to do a graceful shutdown to a 3Node server? How can you shutdown or power off a 3Node server?](#is-it-possible-to-do-a-graceful-shutdown-to-a-3node-server-how-can-you-shutdown-or-power-off-a-3node-server)
+ - [Is it possible to have direct access to Zero-OS's core to force a reboot?](#is-it-possible-to-have-direct-access-to-zero-oss-core-to-force-a-reboot)
+ - [Do I need some port forwarding in my router for each 3Node server?](#do-i-need-some-port-forwarding-in-my-router-for-each-3node-server)
+ - [Are there ways to reduce 3Node servers' noises?](#are-there-ways-to-reduce-3node-servers-noises)
+ - [I built a 3Node out of old hardware. Is it possible that my BIOS or UEFI has improper time and date set as factory default?](#i-built-a-3node-out-of-old-hardware-is-it-possible-that-my-bios-or-uefi-has-improper-time-and-date-set-as-factory-default)
+ - [I have rack servers in my ThreeFold farm. Can I set rack servers vertically?](#i-have-rack-servers-in-my-threefold-farm-can-i-set-rack-servers-vertically)
+ - [Farming and Maintenance](#farming-and-maintenance)
+ - [How can I check if there is utilization on my 3Nodes?](#how-can-i-check-if-there-is-utilization-on-my-3nodes)
+ - [Do I need the Zero-OS bootstrap image drive (USB or CD-DVD) when I reboot, or can I boot Zero-OS from the 3Node main hard drive?](#do-i-need-the-zero-os-bootstrap-image-drive-usb-or-cd-dvd-when-i-reboot-or-can-i-boot-zero-os-from-the-3node-main-hard-drive)
+ - [It's written that my node is using 100% of HRU. What does it mean?](#its-written-that-my-node-is-using-100-of-hru-what-does-it-mean)
+ - [On the ThreeFold Node Finder, I only see half of the virtual cores or threads my 3Node has, what can I do?](#on-the-threefold-node-finder-i-only-see-half-of-the-virtual-cores-or-threads-my-3node-has-what-can-i-do)
+ - [Why are the 3Nodes' resources different on the ThreeFold Node Finder and the ThreeFold Dashboard?](#why-are-the-3nodes-resources-different-on-the-threefold-node-finder-and-the-threefold-dashboard)
+ - [How can I test the health of my disks?](#how-can-i-test-the-health-of-my-disks)
+ - [How can I transfer my 3Node from one farm to another?](#how-can-i-transfer-my-3node-from-one-farm-to-another)
+ - [What do CRU, MRU, HRU and SRU mean on the ThreeFold Node Finder?](#what-do-cru-mru-hru-and-sru-mean-on-the-threefold-node-finder)
+ - [I have more than one ThreeFold 3Node farm, but I want all my 3Nodes on only one farm. How can I put all my 3Nodes on one farm? How can I change the farm ID of my 3Node?](#i-have-more-than-one-threefold-3node-farm-but-i-want-all-my-3nodes-on-only-one-farm-how-can-i-put-all-my-3nodes-on-one-farm-how-can-i-change-the-farm-id-of-my-3node)
+ - [How can I know if my 3Node is online on the Grid?](#how-can-i-know-if-my-3node-is-online-on-the-grid)
+ - [I booted my 3Node and the monitor says it's online and connected to the Grid. But the ThreeFold Node Finder says it is offline? What can I do?](#i-booted-my-3node-and-the-monitor-says-its-online-and-connected-to-the-grid-but-the-threefold-node-finder-says-it-is-offline-what-can-i-do)
+ - [My 3Node does show on the ThreeFold Node Finder, but not on the ThreeFold Dashboard, what can I do?](#my-3node-does-show-on-the-threefold-node-finder-but-not-on-the-threefold-dashboard-what-can-i-do)
+ - [If I upgrade my 3Node, will it increase my rewards?](#if-i-upgrade-my-3node-will-it-increase-my-rewards)
+ - [I booted my 3Node for the first time at the beginning of the month, then I did some upgrade or downgrade, will the ThreeFold Grid recognize the new hardware? Will it still be the same 3Node ID?](#i-booted-my-3node-for-the-first-time-at-the-beginning-of-the-month-then-i-did-some-upgrade-or-downgrade-will-the-threefold-grid-recognize-the-new-hardware-will-it-still-be-the-same-3node-id)
+ - [Is it possible to ask the 3Node to refetch the node information on the monitor?](#is-it-possible-to-ask-the-3node-to-refetch-the-node-information-on-the-monitor)
+ - [When does Zero-OS detect the capacity of a 3Node?](#when-does-zero-os-detect-the-capacity-of-a-3node)
+ - [Where is the 3Node ID stored?](#where-is-the-3node-id-stored)
+ - [Is there a way to backup my node ID in order to restore a 3Node if the disk with the node ID gets corrupted or breaks down?](#is-there-a-way-to-backup-my-node-id-in-order-to-restore-a-3node-if-the-disk-with-the-node-id-gets-corrupted-or-breaks-down)
+ - [If I upgrade my 3Node, does it change the node ID?](#if-i-upgrade-my-3node-does-it-change-the-node-id)
+ - [Does it make sense to recreate my node when the price drops?](#does-it-make-sense-to-recreate-my-node-when-the-price-drops)
+ - [My 3Node lost power momentarily and I had to power it back on manually. Is there a better way to proceed?](#my-3node-lost-power-momentarily-and-i-had-to-power-it-back-on-manually-is-there-a-better-way-to-proceed)
+ - [Do I need to change the battery BIOS?](#do-i-need-to-change-the-battery-bios)
+ - [Do I need to enable UEFI Network Stack?](#do-i-need-to-enable-uefi-network-stack)
+ - [I want redundancy of power for my 3 nodes. I have two PSU on my Dell server. What can I do?](#i-want-redundancy-of-power-for-my-3-nodes-i-have-two-psu-on-my-dell-server-what-can-i-do)
+ - [Why isn't there support for RAID? Does Zero-OS work with RAID?](#why-isnt-there-support-for-raid-does-zero-os-work-with-raid)
+ - [Is there a way to bypass RAID in order for Zero-OS to have bare metals on the system? (No RAID controller in between storage and the Grid.)](#is-there-a-way-to-bypass-raid-in-order-for-zero-os-to-have-bare-metals-on-the-system-no-raid-controller-in-between-storage-and-the-grid)
+ - [I have a 3Node rack server. Is it possible to use a M.2 to SATA adapter in order to put the M.2 SATA disk in the HDD bay (onboard storage)?](#i-have-a-3node-rack-server-is-it-possible-to-use-a-m2-to-sata-adapter-in-order-to-put-the-m2-sata-disk-in-the-hdd-bay-onboard-storage)
+ - [My 3Node uses only PCIe adapters and SSD NVME disks. Do I need the RAID controller on?](#my-3node-uses-only-pcie-adapters-and-ssd-nvme-disks-do-i-need-the-raid-controller-on)
+ - [Can I change the name of my farm on polkadot.js?](#can-i-change-the-name-of-my-farm-on-polkadotjs)
+ - [How can I delete a farm on polkadot.js?](#how-can-i-delete-a-farm-on-polkadotjs)
+ - [I try to delete a node on the TF Dashboard, but it doesn’t work. Is there any other way to proceed that could work?](#i-try-to-delete-a-node-on-the-tf-dashboard-but-it-doesnt-work-is-there-any-other-way-to-proceed-that-could-work)
+ - [My 3Node has 2 ethernet ports in the back, with one written AMT above, what does it mean? Can I use this port to connect my 3Node to the ThreeFold Grid?](#my-3node-has-2-ethernet-ports-in-the-back-with-one-written-amt-above-what-does-it-mean-can-i-use-this-port-to-connect-my-3node-to-the-threefold-grid)
+ - [My 3Node is based on a the hardware Z600, Z620 or Z820, can I run it headless or without a GPU?](#my-3node-is-based-on-a-the-hardware-z600-z620-or-z820-can-i-run-it-headless-or-without-a-gpu)
+ - [Is it possible to add high-level GPU on rack servers to farm more TFT?](#is-it-possible-to-add-high-level-gpu-on-rack-servers-to-farm-more-tft)
+ - [If I change farm, will my node IDs change on my 3Node servers?](#if-i-change-farm-will-my-node-ids-change-on-my-3node-servers)
+ - [Troubleshooting and Error Messages](#troubleshooting-and-error-messages)
+ - [Is it possible to access the Error Screen or Log Screen?](#is-it-possible-to-access-the-error-screen-or-log-screen)
+ - [What does it mean when I see, during the 3Node boot, the message: error = context deadline exceeded?](#what-does-it-mean-when-i-see-during-the-3node-boot-the-message-error--context-deadline-exceeded)
+ - [I try to boot a 3Node, but I get the error: "No Route to Host on Linux". What does it mean?](#i-try-to-boot-a-3node-but-i-get-the-error-no-route-to-host-on-linux-what-does-it-mean)
+ - [How can I fix the error: "Network configuration succeed but Zero-OS kernel could not be downloaded" when booting a 3Node?](#how-can-i-fix-the-error-network-configuration-succeed-but-zero-os-kernel-could-not-be-downloaded-when-booting-a-3node)
+ - [Using SAS disks, I get the error; "No ssd found, failed to register". What can I do to fix this?](#using-sas-disks-i-get-the-error-no-ssd-found-failed-to-register-what-can-i-do-to-fix-this)
+ - [When booting a 3Node, how to fix the error: "no disks: registration failed"?](#when-booting-a-3node-how-to-fix-the-error-no-disks-registration-failed)
+ - [My SSD is sometimes detected as HDD by Zero-OS when there is a reboot. Is there a fix or a way to test the SSD disk?](#my-ssd-is-sometimes-detected-as-hdd-by-zero-os-when-there-is-a-reboot-is-there-a-fix-or-a-way-to-test-the-ssd-disk)
+ - [When booting a 3Node, I get the message: failed to register node: failed to create node: failed to submit extrinsic: Invalid Transaction: registration failed. What could fix this?](#when-booting-a-3node-i-get-the-message-failed-to-register-node-failed-to-create-node-failed-to-submit-extrinsic-invalid-transaction-registration-failed-what-could-fix-this)
+ - [I try to boot a 3Node, but I get the message no route with default gateway found. What does it mean?](#i-try-to-boot-a-3node-but-i-get-the-message-no-route-with-default-gateway-found-what-does-it-mean)
+ - [I have trouble connecting the 3Node to the Grid with a 10GB NIC card. What can I do?](#i-have-trouble-connecting-the-3node-to-the-grid-with-a-10gb-nic-card-what-can-i-do)
+ - [I switch the ethernet cable to a different port when my 3Node was running. Internet connection is lost. What can I do?](#i-switch-the-ethernet-cable-to-a-different-port-when-my-3node-was-running-internet-connection-is-lost-what-can-i-do)
+ - [I get the error Certificate is not yet valid when booting my 3Node server, what can I do?](#i--get-the-error-certificate-is-not-yet-valid-when-booting-my-3node-server-what-can-i-do)
+ - [When running wipefs to wipe my disks on Linux, I get either of the following errors: "syntax error near unexpected token" or "Probing Initialized Failed". Is there a fix?](#when-running-wipefs-to-wipe-my-disks-on-linux-i-get-either-of-the-following-errors-syntax-error-near-unexpected-token-or-probing-initialized-failed-is-there-a-fix)
+ - [I did a format on my SSD disk, but Zero-OS still does not recognize them. What's wrong?](#i-did-a-format-on-my-ssd-disk-but-zero-os-still-does-not-recognize-them-whats-wrong)
+ - [I have a Dell Rx10 server (R610, 710, 910). When I boot Zero-OS I get the message Probing EDD and the 3Node doesn't boot from there. What can I do?](#i-have-a-dell-rx10-server-r610-710-910-when-i-boot-zero-os-i-get-the-message-probing-edd-and-the-3node-doesnt-boot-from-there-what-can-i-do)
+ - [My 3Node doesn't boot properly without a monitor plugged in. What can I do?](#my-3node-doesnt-boot-properly-without-a-monitor-plugged-in-what-can-i-do)
+ - [My 3Node is running on the Grid, but when I plugged in the monitor, it states: Disabling IR #16. Is there a problem?](#my-3node-is-running-on-the-grid-but-when-i-plugged-in-the-monitor-it-states-disabling-ir-16-is-there-a-problem)
+ - [My 3Node won't boot without disabling the Secure Boot option, is it safe?](#my-3node-wont-boot-without-disabling-the-secure-boot-option-is-it-safe)
+ - [When I tried to boot my 3Node, at some point the screen went black, with or without a blinking hyphen or dash. What could cause this and what could I do to resolve the issue?](#when-i-tried-to-boot-my-3node-at-some-point-the-screen-went-black-with-or-without-a-blinking-hyphen-or-dash-what-could-cause-this-and-what-could-i-do-to-resolve-the-issue)
+ - [My 3Nodes go offline after a modem reboot. Is there a way to prevent this?](#my-3nodes-go-offline-after-a-modem-reboot-is-there-a-way-to-prevent-this)
+ - [When I boot my 3Node, it reaches the Welcome to Zero-OS window, but it doesn't boot properly and there's an error message: failed to load object : type substrate..., what can I do?](#when-i-boot-my-3node-it-reaches-the-welcome-to-zero-os-window-but-it-doesnt-boot-properly-and-theres-an-error-message-failed-to-load-object--type-substrate-what-can-i-do)
+ - [When I try to access iDRAC on a web browswer, even with protected mode off, I get the error The webpage cannot be found, what can I do?](#when-i-try-to-access-idrac-on-a-web-browswer-even-with-protected-mode-off-i-get-the-error-the-webpage-cannot-be-found-what-can-i-do)
+ - [When booting the 3Node, I get the error Network interface detected but autoconfiguration failed. What can I do?](#when-booting-the-3node-i-get-the-error-network-interface-detected-but-autoconfiguration-failed-what-can-i-do)
+ - [When I boot my Dell server, I get the message: All of the disks from your previous configuration are gone... Press any key to continue or 'C' to load the configuration utility. What can I do?](#when-i-boot-my-dell-server-i-get-the-message-all-of-the-disks-from-your-previous-configuration-are-gone-press-any-key-to-continue-or-c-to-load-the-configuration-utility-what-can-i-do)
+ - [I have a Dell R620. In Zero-OS, I get the failure message No network card found and then the 3Node reebots after few seconds. The same happens for every LAN input. What can I do?](#i-have-a-dell-r620-in-zero-os-i-get-the-failure-message-no-network-card-found-and-then-the-3node-reebots-after-few-seconds-the-same-happens-for-every-lan-input-what-can-i-do)
+ - [I am using freeDos to crossflash my raid controller on a Dell server, but I can't see the RAID controller with the Command Info. What can I do?](#i-am-using-freedos-to-crossflash-my-raid-controller-on-a-dell-server-but-i-cant-see-the-raid-controller-with-the-command-info-what-can-i-do)
+ - [Can I use a VGA to HDMI adaptor to connect a TV screen or monitor to the 3Node? I tried to boot a 3Node with a VGA to HDMI adaptor but the boot fails, what can I do?](#can-i-use-a-vga-to-hdmi-adaptor-to-connect-a-tv-screen-or-monitor-to-the-3node-i-tried-to-boot-a-3node-with-a-vga-to-hdmi-adaptor-but-the-boot-fails-what-can-i-do)
+ - [When I try to boot my 3Node, the fans start spinning fast with a loud noise and the screen is black. What can I do to resolve this?](#when-i-try-to-boot-my-3node-the-fans-start-spinning-fast-with-a-loud-noise-and-the-screen-is-black-what-can-i-do-to-resolve-this)
+ - [When booting Zero-OS with IPV6 configurations, I get the errors (1) dial tcp: address IPV6-address too many columns in address and (2) no pools matches key: not routable. What can I do to fix this issue?](#when-booting-zero-os-with-ipv6-configurations-i-get-the-errors-1-dial-tcp-address-ipv6-address-too-many-columns-in-address-and-2-no-pools-matches-key-not-routable-what-can-i-do-to-fix-this-issue)
+ - [When booting a 3Node, Zero-OS downloads fine, but then I get the message: error no route with default gateway found, and the message: info check if interface has a cable plugged in. What could fix this?](#when-booting-a-3node-zero-os-downloads-fine-but-then-i-get-the-message-error-no-route-with-default-gateway-found-and-the-message-info-check-if-interface-has-a-cable-plugged-in-what-could-fix-this)
+ - [How can I update Dell and HP servers to Intel E5-2600v2, E5-2400v2 and E5-4600v2, when applicable?](#how-can-i-update-dell-and-hp-servers-to-intel-e5-2600v2-e5-2400v2-and-e5-4600v2-when-applicable)
+ - [How can I update the firmware and driver of a Dell PowerEdge server?](#how-can-i-update-the-firmware-and-driver-of-a-dell-poweredge-server)
+ - [When I boot a 3Node in UEFI mode, it gets stuck at: Initializing Network Device, is there a way to fix this?](#when-i-boot-a-3node-in-uefi-mode-it-gets-stuck-at-initializing-network-device-is-there-a-way-to-fix-this)
+ - [When I boot my 3Node, it gets stuck during the Zero-OS download. It never reaches 100%. What can I do to fix this issue?](#when-i-boot-my-3node-it-gets-stuck-during-the-zero-os-download-it-never-reaches-100-what-can-i-do-to-fix-this-issue)
+ - [When booting a 3Node, I get the error=“context deadline exceeded” module=network error=failed to initialize rmb api failed to initialized admin mw: failed to get farm: farm not found: object not found. What can I do to fix this issue?](#when-booting-a-3node-i-get-the-errorcontext-deadline-exceeded-modulenetwork-errorfailed-to-initialize-rmb-api-failed-to-initialized-admin-mw-failed-to-get-farm-farm-not-found-object-not-found-what-can-i-do-to-fix-this-issue)
+ - [ThreeFold Grid and Data](#threefold-grid-and-data)
+ - [How is the farming minting reward calculated? Is the Grid always monitoring my 3Node?](#how-is-the-farming-minting-reward-calculated-is-the-grid-always-monitoring-my-3node)
+ - [How does communication happen on the ThreeFold Grid at the 3Node's level?](#how-does-communication-happen-on-the-threefold-grid-at-the-3nodes-level)
+ - [What is the ThreeFold Node Status bot Telegram link?](#what-is-the-threefold-node-status-bot-telegram-link)
+ - [How does the ThreeFold Node Status bot work? How can I use the ThreeFold Node Status bot to verify if my 3Node is online?](#how-does-the-threefold-node-status-bot-work-how-can-i-use-the-threefold-node-status-bot-to-verify-if-my-3node-is-online)
+ - [How does the Telegram Status Bot get information from my 3Node? My 3Node is online on the ThreeFold Node Finder, but offline on the Telegram Status Bot, is this normal?](#how-does-the-telegram-status-bot-get-information-from-my-3node-my-3node-is-online-on-the-threefold-node-finder-but-offline-on-the-telegram-status-bot-is-this-normal)
+ - [I noticed that when I reboot my 3Node, the uptime counter on the ThreeFold Node Finder goes back to zero. Does it mean I lose uptime and the uptime start over again when I reboot the 3Node?](#i-noticed-that-when-i-reboot-my-3node-the-uptime-counter-on-the-threefold-node-finder-goes-back-to-zero-does-it-mean-i-lose-uptime-and-the-uptime-start-over-again-when-i-reboot-the-3node)
+ - [One of my nodes is showing the wrong location. Any problem with that?](#one-of-my-nodes-is-showing-the-wrong-location-any-problem-with-that)
+ - [Memory](#memory)
+ - [Can I use different type of RAM for the same 3Node?](#can-i-use-different-type-of-ram-for-the-same-3node)
+ - [How can I know if the memory I am buying is correct for my specific hardware?](#how-can-i-know-if-the-memory-i-am-buying-is-correct-for-my-specific-hardware)
+ - [What do the terms RDIMM, LDIMM, UDIMM, LRDIMM, FBDIMM mean when it comes to RAM memory sticks?](#what-do-the-terms-rdimm-ldimm-udimm-lrdimm-fbdimm-mean-when-it-comes-to-ram-memory-sticks)
+ - [What is the difference between ECC and non-ECC memory?](#what-is-the-difference-between-ecc-and-non-ecc-memory)
+ - [How can I change the RAM memory sticks on my 3Nodes? How can I achieve dual channel configuration with sticks of RAM?](#how-can-i-change-the-ram-memory-sticks-on-my-3nodes-how-can-i-achieve-dual-channel-configuration-with-sticks-of-ram)
+ - [What does RAM mean?](#what-does-ram-mean)
+ - [What does DIMM mean when it comes to RAM sticks?](#what-does-dimm-mean-when-it-comes-to-ram-sticks)
+ - [I have 24 DIMMS ram slots on my server. Can I use them all?](#i-have-24-dimms-ram-slots-on-my-server-can-i-use-them-all)
+- [Ask a Question to the ThreeFold Community](#ask-a-question-to-the-threefold-community)
+
+***
+
+# GENERAL FAQ
+
+## Basic Facts
+
+### What is the the ThreeFold blockchain?
+
+ThreeFold blockchain is the layer 0 infrastructure for an open source peer-to-peer (P2P) Internet owned by humanity.
+
+
+
+### What is the architecture of the ThreeFold Grid in simple terms?
+
+Essentially, the ThreeFold Grid is composed of the people using it, the 3Node servers offering compute, storage and network resources, and the TF Chain, which is the blockchain of ThreeFold.
+
+Middlewares are also used, such as GraphQL and GrixProxy, to get and organize data from the ThreeFold Chain. They help to make data available and to manage load.
+
+3Nodes store workloads data and can report on their state to the TF Grid and to middlewares.
+
+
+### What is the difference between Internet capacity and connectivity? Does ThreeFold replace my Internet service provider (ISP)?
+
+In simple terms, the Internet is composed of both capacity and connectivity. Capacity is where the data and resources are being handled, for example in servers. Connectivity is the infrastructure that transfer data and resources between servers. The latter is linked to the typical Internet service provider (ISP).
+
+ThreeFold’s technology enables distributed capacity generation, but ThreeFold doesn’t deal in connectivity.
+3nodes offer Internet capacity, but farmers still rely on connectivity provider like the usual Internet service provider (ISP).
+
+### What are the priorities of ThreeFold (the Three P of ThreeFold)? ThreeFold is a Planet first project, what does it mean?
+
+ThreeFold is working for the Planet, the People and Profit, in this very order of importance. Planet comes first as it is our home to us all. A humane enterprise always has people before profit, and serious entrepreneurs know profit cannot be left out of the equation of a thriving project.
+
+
+### I want to help build the new Internet. How can I become a ThreeFold certified 3node partner?
+
+Apply [here](https://marketplace.3node.global/index.php?dispatch=companies.apply_for_vendor) to become a ThreeFold certified 3node partner.
+
+
+
+### How can I create a twin on the TF Grid?
+
+There are 2 ways to create a twin:
+
+You can create a twin via the [ThreeFold Dashboard](../dashboard/wallet_connector.md).
+
+You can also create a twin via the ThreeFold Connect app. Indeed, a twin is automatically generated while creating a farm. Note that, in this case, the twin will be created on mainnet.
+
+
+## ThreeFold Communication
+
+### Is there a ThreeFold app for mobile?
+
+Yes! ThreeFold Connect App (TF Connect App) is available for [Android](https://play.google.com/store/apps/details?id=org.jimber.threebotlogin) and [iOS](https://apps.apple.com/us/app/threefold-connect/id1459845885).
+
+You can use this app to create a ThreeFold ID, a ThreeFold Wallet and also a ThreeFold Farm to link all your 3nodes.
+The ThreeFold Connect Wallet, with its Stellar payout address, can be used for transactions as well as to receive farming rewards.
+The News section gives you the latest information on the fast ThreeFold development and growth.
+
+
+### I want to reach the ThreeFold community. What are ThreeFold social links?
+
+You can find links to the ThreeFold community at the bottom of the main page of the [ThreeFold website](https://www.threefold.io/).
+
+
+### Could we reach out someone for publishing research work on ThreeFold?
+
+You can send an email to info@threefold.io for publishing research on ThreeFold.
+
+
+### Who can I write to for a proposal? Where can I send a proposal email for a new partnership opportunity with ThreeFold?
+
+You can mail your proposal to info@threefold.io or write about your proposal on the [ThreeFold Forum](http://forum.threefold.io/).
+
+
+
+### How can I track and follow the progress and development of ThreeFold?
+
+There are two main places where you can track the progress of ThreeFold. ThreeFold is open source and its developments can be easily tracked on Github.
+
+* You can read about the ongoing ThreeFold Tech projects [here](https://github.com/orgs/threefoldtech/projects).
+* You can read about the ongoing ThreeFold Foundation projects [here](https://github.com/orgs/threefoldfoundation/projects?query=is%3Aopen).
+
+
+
+### Why do some forum posts need to be approved?
+
+The configurations of the ThreeFold forum make so that some posts need approval, depending on where they are posted. Like posting from meta-sections, and some specific sections.
+
+Also, note that certain posts can automatically get flagged for moderation based on their content.
+
+
+
+## The Technology of ThreeFold
+
+### What is a 3Node?
+
+It is essentially a single server that makes up a larger network of servers which together form the ThreeFold Grid. Essentially any modern computer can be turned into a 3node (DIY Farming) and you can [buy plug and play 3nodes](https://marketplace.3node.global/index.php) as state of the art modern computer.
+
+
+
+### What is the difference between a 3node and a ThreeFold farm?
+
+A 3node is a single server connected to the Grid. Each 3node is linked to a farm. A farm can be composed of multiple 3nodes.
+
+
+### What is Zero-OS from ThreeFold?
+
+Zero-OS is a stateless and lightweight operating system designed to host anything that runs on Linux, in a decentralized way. Once installed, Zero-OS locks the hardware and dedicates its capacity to the People’s Internet via the ThreeFold Blockchain.
+
+
+
+### ThreeFold uses Quantum Safe Storage technology, what does it mean?
+
+Quantum computers are theoretically capable of doing huge calculations in a short period of time. By this fact alone, it is a great potential threat to future online safety. ThreeFold solves this future problem before it even becomes a reality. Indeed, Zero-os can compress, encrypt, and disperse data across the Grid.
+
+
+
+### Quantum Safe File System (QSFS) allows for part of the storage to go down and it can self repair, however it’s still attached to a single VM and a single point of failure. Can a QSFS instance be reattached to another VM to recover it?
+
+QSFS is built from storage devices which are distributed and decentralized.
+
+The storage engine is a software running on a VM that can run everywhere.
+If the storage engine needs to run on a different VM the config needs to pushed to the new VM.
+In short, yes Quantum safe file system (QSFS) can be recovered on a different VM. It is not automated yet on Zero-OS. A video tutorial will be shared soon.
+
+
+
+
+### Where does the ThreeFold Explorer take its data from?
+
+The ThreeFold Explorer takes its data from this website: [https://gridproxy.grid.tf/](https://gridproxy.grid.tf/).
+
+To explore Grid Proxy, you can use Swagger: [https://gridproxy.grid.tf/swagger/index.html](https://gridproxy.grid.tf/swagger/index.html). You will then be able to query the TF Grid and extract data.
+
+See the next Q&A for more information on Swagger.
+
+
+
+### How can I use the Gridproxy to query information on the TF Grid?
+
+You can go to the Gridproxy Swagger index: [https://gridproxy.grid.tf/swagger/index.html](https://gridproxy.grid.tf/swagger/index.html).
+
+There you can query information such as information on a 3node.
+
+For example, asking the Gridproxy for the nodeID 466, you get the following URL: `https://gridproxy.grid.tf/nodes/466`.
+
+To get specific information, you can add parameters, for example: `https://gridproxy.grid.tf/nodes/466/status`.
+
+When you know the URL representation of the query, you can simply use the URL directly on a web browser.
+
+
+
+### How can I see the stats of the ThreeFold Grid?
+
+You can go to think link: [https://stats.grid.tf/](https://stats.grid.tf/) to see the stats of the ThreeFold Grid.
+
+You can also check this [thread](https://forum.threefold.io/t/grid-stats-new-nodes-utilization-overview/3291/) on the ThreeFold forum.
+
+
+### What is the difference between a seed phrase (mnemonics) and an HEX secret?
+
+A seed phrase (also called mnemonics) is a set of words from a carefully selected pool that can be used to derive cryptographic secrets. A HEX secret is a more direct representation of such secret that the computer uses. In the case of an HEX secret, there is no extra information present to form complete words of the seed phrase.
+
+In typical usage, multiple secrets can be derived from a seed phrase via a one-way operation. The secret and its derived public key are sufficient to do any cryptographic operation like signing transactions or encrypting data, but it can't be used to get back the words of the seed phrase.
+
+
+
+## Buying and Transacting TFT
+
+
+### How long does it take when you use the BSC-Stellar Bridge?
+
+The bridge will process deposits/withdrawals within 48 hours.
+
+
+
+### On my website, users can donate TFT on the Stellar Chain. Is there a way for users on my website to easily track the total sum of TFT donated?
+
+There is a simple way to do this. The [Stellar Explorer](https://stellar.expert/explorer/public) has an embeddable widget that you can insert on any website, including WordPress.
+
+Simply go to the account you’re interested in showing the balance of, look for “Balance History”, select TFT, and finally click the small icon next to the heading to reveal the embed code. In your WordPress page editor, in HTML mode, paste the embed code.
+
+
+## TF Connect App, TF Dashboard, GraphQL, Grix Proxy and Polkadot Substrate
+
+### Is there a way to create or import another wallet in TF Connect App?
+
+The TF Connect App supports Stellar and TF Chain wallets. The app by default can create one wallet. To add any number of additional wallets, you must create a wallet on Stellar or TF Chain and then import it with the import function.
+
+
+
+### I created a farm on the TF Chain. On the TF Connect App Farmer Migration section, my farm is under Other v3 farms, is this normal?
+
+Yes this is normal. Farms created on TF Chain instead of the TF Connect App will appear in *Other v3 farms*.
+
+
+
+### I am trying to access my wallet in the ThreeFold Connect App. It worked fine before, but now I just get a white screen. What does it mean and what can I do?
+
+On the TF Connect App, when you get a white screen, it means that there is a connection issue. It can help to try other networks; maybe try switching between ethernet cable or wifi. Or you can also try it later when the connection might be more stable.
+
+
+
+### When I open the ThreeFold Connect App, I get the error: Error in initialization in Flagsmith. How can I fix this issue?
+
+To fix this Flagsmith error message on the ThreeFold Connect app, you can try the following methods:
+
+* Check your internet connection
+* Update your phone current operating system (OS) version
+* Update the date and time on your phone
+
+
+### Apart form the ThreeFold Connect App Wallet, how can I check my TFT balance?
+
+You can go on [Stellar.Expert](https://stellar.expert). With your wallet address, you will be able to see your transactions and wallet details.
+
+
+
+### Is it possible to export the transaction history of a wallet to a CSV file?
+
+Yes, every blockchain has an explorer function and these explorer functions allow you to see transactions and export them. TFT is on 2 chains at the moment: Stellar and Polkadot.
+
+For Stellar based TFT’s there is an explorer here: https://stellar.expert/explorer/public. Enter you wallet address in the top left search box, and after pressing enter you should see every transaction related to your account.
+
+If you are not deploying/doing things on the TF Grid (dev, test or mainnet) you will not have transferred any tokens to the TF Chain, therefore all your tokens/wallets will be on the Stellar Chain.
+
+
+### How can I use GraphQl to find information on the ThreeFold Grid?
+
+To find information on the ThreeFold Grid with GraphQL, go to this [link](https://graphql.grid.tf/graphql). On the left menu, choose the parameters you want to search and write the necessary information, if needed, then click on the Play button in the middle section, at the top.
+
+Here's an example of a query, where we want to find all the farms containing "duck" in their name.
+
+query MyQuery {
+ farms(where: {name_contains: "duck"}) {
+ name
+ farmID
+ }
+}
+
+This code can be written automatically if you simply select the proper parameters in the left menu.
+For the previous example, we had to click on "farms", then "where", and then "name_contains". After clicking on "name_contains", you need to add the words you are looking for, in this example we had "duck". Further down the menu, we simply had to click on "farmID" and "name", and then click the Play button. The results of the query appear on the right screen.
+
+
+
+### What are the different links to ThreeFold's Graph QL depending on the network?
+
+The links for the Development, Test and Main Networks are the following:
+
+* Main Net Graph QL
+ * [https://graphql.grid.tf/graphql](https://graphql.grid.tf/graphql)
+* Test Net Graph QL
+ * [https://graphql.test.grid.tf/graphql](https://graphql.test.grid.tf/graphql)
+* Dev Net Graph QL
+ * [https://graphql.dev.grid.tf/graphql](https://graphql.dev.grid.tf/graphql)
+
+
+
+### How can I find 3Nodes with IPv6 addresses?
+
+You can use [GraphQL](https://graphql.grid.tf/graphql) for such queries.
+
+Use the following code to search for 3Nodes with IPv6 addresses.
+Enter the following code on the middle window and click on the "Play" button.
+
+```
+query MyQuery {
+ publicConfigs {
+ ipv6
+ node {
+ nodeID
+ }
+ }
+}
+```
+
+The 3nodes with IPv6 addresses will appear on the right window.
+
+For more information on how to use Graph QL, read [this Q&A](#how-can-i-use-graphql-to-find-information-on-the-threefold-grid).
+
+
+
+### How can I use GraphQL to see contracts on my 3Nodes?
+
+Go to [ThreeFold's GraphQL](https://graphql.grid.tf/graphql) and write the following query:
+
+```
+query MyQuery {
+ nodeContracts(where: {state_eq: Created, nodeID_eq: 42}) {
+ resourcesUsed {
+ cru
+ hru
+ mru
+ sru
+ }
+ contractID
+ twinID
+ nodeID
+ }
+}
+
+```
+
+This will show you contracts on the 3Node as well as resources used. You can play with the different parameters.
+
+How can I see the farm associated with a node?
+
+```
+query MyQuery {
+ nodes(where: {nodeID_eq: 57}) {
+ farmID
+ nodeID
+ }
+}
+```
+
+
+
+### How can I use Grid Proxy to find information on the ThreeFold Grid and 3Nodes?
+
+To find information on the ThreeFold Grid with GraphQL, you need to write this URL: https://gridproxy.grid.tf/, followed by your specific query. Here's an example if we wanted to see all the available farm on the TF Grid that has "duck" in its name:
+
+https://gridproxy.grid.tf/farms?name_contains=duck
+
+The Grid Proxy is appropriate for high volume application.
+You can find the parameters to be written in the URL when visiting the [GraphQL explorer](https://graphql.grid.tf/graphql).
+
+
+
+### Who is hosting GraphQL and Grid Proxy on the ThreeFold Grid?
+
+GraphQL and Grid Proxy are hosted by ThreeFold for everyone to use.
+
+Note that it is also possible to run your own instance of those tools.
+
+
+
+### What is the difference between uptime, status and power state?
+
+There are three distinctly named endpoints or fields that exist in the back end systems:
+
+* Uptime
+ * number of seconds the node was up, as of it's last uptime report. This is the same on GraphQL and Grid Proxy.
+* Status
+ * this is a field that only exists on the Grid Proxy, which corresponds to whether the node sent an uptime report within the last 40 minutes.
+* Power state
+ * this is a field that only exists on GraphQL, and it's the self reported power state of the node. This only goes to "down" if the node shut itself down at request of the Farmerbot.
+
+
+
+### I do not remember the name (ThreeFold 3bot ID) associated with my seed phrase on the ThreeFold Connect app. Can I recover my TF Connect app account with only the seed phrase and not the name (3bot ID) associated with it?
+
+If you forgot the name associated with your seed phrase on the TF Connect app, you can always create a new identity (ThreeFold 3bot ID) and import your wallet using the old seed phrase.
+
+Since the Connect App is also used for identity and authentication, you need both the name (3bot ID) and seed phrase to fully recover your account. The wallet is only linked to the seed phrase and not the name (3bot ID).
+
+
+# USERS FAQ
+
+## TF Grid Functionalities
+
+
+### What are the type of storage available on TF Grid?
+
+There’s two type of storage that van de used on the TF Grid.
+
+1. VM which has a virtual disk. The virtual disk is a straightforward volume on a local hard disk. Everything stored on this virtual disk is stored only on this virtual (and thus physical) disk. Delete the VM and the virtual disk and the content is gone.
+2. Quantum safe storage. Quantum safe storage uses a “Storage Engine” that parts, compresses, encrypts and then mathematically describes the data.
+
+
+## Deployments on the ThreeFold Grid
+
+
+### Does the ThreeFold Grid charge the total resources rented or it only charges the resources used during deployment?
+
+Billing is based on how many resources you reserve, not how much you use them. For this reason, it can be a good idea to deploy the minimim resources needed per project.
+
+
+
+### Do I pay for Internet traffic while deploying workloads on IPv4, IPv6 or Planetary Network?
+
+You do pay for internet traffic while deploying on the ThreeFold Grid. It is calculated during deployment and paid with ThreeFold tokens (TFT).
+
+Note that the private overlay network traffic is not billed.
+
+
+
+### What is the monthly cost for an IPv4 or an IPv6 public address on the ThreeFold Grid?
+
+The cost for an IPv4 public address is around 3$/month (USD).
+
+For an IPv6 address, there is no cost.
+
+
+
+### What are the differences between a container, a micro virtual machine and a full virtual machine (VM)?
+
+The following is a list of certain features related to containers as well as full and micro virtual machines.
+
+* Container
+ * generally designed to run a single application
+ * doesn't need to include a full operating system
+ * relies on the kernel of the host system, no hypervisor needed
+ * isolated from the rest of the host system for security and can be limited in resources used
+ * examples: on the [Playground](https://playground.grid.tf/), we have [Kubernetes](https://library.threefold.me/info/manual/#/manual__weblets_k8s?id=kubernetes) and [Caprover](https://library.threefold.me/info/manual/#/manual__weblets_caprover?id=caprover), which are both environments that host containers
+* Micro VM
+ * a container image promoted to run as a VM by pairing with a generic kernel
+ * more isolated than a container, thus more secure
+ * generally lighter than a full VM
+ * can be created from any Docker container image by uploading it to the [TF Hub](https://hub.grid.tf/)
+ * examples: on the [Playground](https://playground.grid.tf/), we have Ubuntu 20.04, Alpine-3, CentOS-8 and more.
+* Full VM
+ * contains a complete operating system including kernel
+ * capable of anything that can be done with a Linux server
+ * compatible with any guides and tutorials written for the same version of the distribution they are running
+ * normally contains systemd, unlike containers which normally do not
+ * can load kernel modules or replace the kernel entirely, so has best compatibility
+ * generally heavier than micro VM
+ * examples: on the [Playground](https://playground.grid.tf/), we have Ubuntu 18.04, 20.04, 22.04 and more
+
+Note that you can run Kubernetes on a micro VM and you can run a very minimal operating system in a full VM. There are many possibilities when using those technologies.
+
+
+
+### What is a 3Node gateway? How can I configure a 3Node as a gateway node?
+
+A 3Node becomes a gateway when a ThreeFold farmer adds a public IP address to the node itself on the [ThreeFold Dashboard](https://dashboard.grid.tf/). In doing so, the IP address is then handed over to the base operating system of the node itself. The IP address can then be used in the overall functions of the TF Grid.
+
+Note that this process differs from when an IP address that has been added to a farm is deployed with a workload in order for that workload to be accessible on the Internet.
+
+To configure a 3Node as a gateway node, you need a public IP block from your internet service provider (ISP).
+
+You can configure a 3Node as a gateway node on the [TF mainnet](https://dashboard.grid.tf/), [TF testnet](https://dashboard.test.grid.tf/) and [TF devnet](https://dashboard.dev.grid.tf/). You thus need to choose the correct TF Dashboard link (main, test, dev).
+
+To configure a 3Node as a gateway node, follow these steps:
+
+* Configure your DNS records
+ * Type: A
+ * Name:
+ * Value:
+ * Type: NS
+ * Name: _acme-challenge.
+ * Value: .
+ * Type: CNAME
+ * Name: *..
+ * Value:
+ * Type: AAAA
+ * Name:
+ * Value:
+* Configure your 3Node parameters on the TF Dashboard
+ * Go to the [ThreeFold Dashboard](https://dashboard.grid.tf/)
+ * Go to the section **Portal**
+ * Go to the subsection **Farms**
+ * Choose the 3Node you want to turn into a gateway node and click on **Actions** (Add a public config) on the right
+ * Enter the necessary information and click **Save**
+ * IPV4: Enter the IPv4 address of your public IP block
+ * Gateway: Enter the gateway of your public IP block
+ * IPV6: Enter the IPv6 address of your public IP block
+ * Gateway IPV6: Enter the gateway of your public IP block
+ * Domain: .
+
+Once this is done, you should see the IPv4 and IPv6 addresses in the section **PUB** of your 3Node screen.
+
+To learn more about this process, [watch this great video](https://youtu.be/axvKipK7MQM).
+
+
+
+### When connecting remotely with SSH, I get the following error: "WARNING: REMOTE HOST IDENTIFICATION HAS CHANGED!...". What can I do to fix this?
+
+If you've already done an SSH connection on your computer, the issue is most probably that the "host key has just been changed". To fix this, try one of those two solutions:
+
+* Linux and MAC:
+ * ```
+ sudo rm ~/.ssh/known_hosts
+ ```
+* Windows:
+ * ```
+ rm ~/.ssh/known_hosts
+ ```
+
+To be more specific, you can remove the probematic host:
+
+* Windows, Linux and MAC:
+ * ```
+ ssh-keygen -R
+ ```
+
+Once this is done, you should be able to SSH into your 3Node deployment.
+
+
+
+### How can I remove one host from known_hosts?
+
+You can write the following command
+```
+ssh-keygen -R hostname
+```
+
+Where hostname would be the IPv4 or IPv6 address for example.
+In the case of the ThreeFold Grid, it can also be the Planetary Network address.
+
+
+
+### How can I add ThreeFold peers in the Yggdrasil configuration file?
+
+In the file /etc/yggdrasil.conf, write the following in the "Peers" section:
+
+```
+ Peers: [
+ # Threefold Lochrist
+ tcp://gent01.grid.tf:9943
+ tcp://gent02.grid.tf:9943
+ tcp://gent03.grid.tf:9943
+ tcp://gent04.grid.tf:9943
+ tcp://gent01.test.grid.tf:9943
+ tcp://gent02.test.grid.tf:9943
+ tcp://gent01.dev.grid.tf:9943
+ tcp://gent02.dev.grid.tf:9943
+ # GreenEdge
+ tcp://gw291.vienna1.greenedgecloud.com:9943
+ tcp://gw293.vienna1.greenedgecloud.com:9943
+ tcp://gw294.vienna1.greenedgecloud.com:9943
+ tcp://gw297.vienna1.greenedgecloud.com:9943
+ tcp://gw298.vienna1.greenedgecloud.com:9943
+ tcp://gw299.vienna2.greenedgecloud.com:9943
+ tcp://gw300.vienna2.greenedgecloud.com:9943
+ tcp://gw304.vienna2.greenedgecloud.com:9943
+ tcp://gw306.vienna2.greenedgecloud.com:9943
+ tcp://gw307.vienna2.greenedgecloud.com:9943
+ tcp://gw309.vienna2.greenedgecloud.com:9943
+ tcp://gw313.vienna2.greenedgecloud.com:9943
+ tcp://gw324.salzburg1.greenedgecloud.com:9943
+ tcp://gw326.salzburg1.greenedgecloud.com:9943
+ tcp://gw327.salzburg1.greenedgecloud.com:9943
+ tcp://gw328.salzburg1.greenedgecloud.com:9943
+ tcp://gw330.salzburg1.greenedgecloud.com:9943
+ tcp://gw331.salzburg1.greenedgecloud.com:9943
+ tcp://gw333.salzburg1.greenedgecloud.com:9943
+ tcp://gw422.vienna2.greenedgecloud.com:9943
+ tcp://gw423.vienna2.greenedgecloud.com:9943
+ tcp://gw424.vienna2.greenedgecloud.com:9943
+ tcp://gw425.vienna2.greenedgecloud.com:9943
+ ]
+```
+
+
+
+### How can I see Yggdrasil/Planetary Network's peers?
+
+On MAC and Linux, write the following line in the terminal:
+
+```
+/etc/yggdrasil.conf
+```
+
+On Windows, one of the two following lines should work:
+
+```
+%ALLUSERSPROFILE%\Yggdrasil\yggdrasil.conf
+```
+
+```
+C:\ProgramData\Yggdrasil\
+```
+
+These are the general location of Yggdrasil. It can change depending on how you installed Yggdrasil.
+
+
+
+### How can I ping an Yggdrasil IP or IPv6 address?
+
+Usually, using the command `ping` works.
+
+If the typical `ping` doesn't work, try this instead: `ping6`
+
+For example, the following will send 2 pings:
+
+```
+ping6 -c 2 yggdrasil_address
+```
+
+
+
+
+### Is there a way to test if I am properly connected to the Yggdrasil network (Planetary Network)?
+
+To check if you are properly connected to the Yggdrasil network, try reaching this website:
+
+```
+http://[319:3cf0:dd1d:47b9:20c:29ff:fe2c:39be]/
+```
+
+If you can reach this website, it means that you are properly connected.
+
+For more information on how to connect to Yggrasil (and the Planetary Network), read [this guide](../system_administrators/getstarted/ssh_guide/ssh_guide.md).
+
+
+
+
+### How can I change the username of my SSH key?
+
+In the terminal, write the following line:
+
+```
+ssh-keygen -C newname
+```
+
+Make sure to replace "newname" by the name you want.
+
+
+### What is ThreeFold's stance on sharded workload? Will ThreeFold embrace and move towards distributed data chunks or stay with the one deployment, one node model?
+
+The ThreeFold Grid is basically agnostic when it comes to how you structure your deployment.
+
+If you want to put all your storage and compute on one node and lose everything if it goes down, you can do so.
+
+If you want a highly distributed and fault tolerant system with high availability where data is never lost, you can also do so. You could build an architecture with single nodes running single workloads as the building blocks.
+
+Containerized micro service architectures running on e.g. Kubernetes are basically the way that compute is being "sharded" already in the mainstream of IT, especially at large scales. Such applications fit well on the ThreeFold Grid today, since they account for the possibility that individual nodes may fail. However, many applications aren't built this way and it takes work to adapt them.
+
+Self driving and self healing IT is one of the core concepts of ThreeFold. What we've built so far is an excellent foundation for making this reality. Some additional features might come into TF Chain and the network itself to enable such features. That being said, a lot is already possible using the existing system. It's just a matter of gluing things together in the right way.
+
+
+
+## Tutorials and Guides
+
+
+### What is the minimum amount of TFT to deploy a Presearch node? How can I get a TFT discount when I deploy a Presearch node?
+
+The minimum amount of TFT that needs to be in your ThreeFold Profile before you can deploy a Presearch node is 2 TFT. But this would not last very long.
+
+To benefit from the biggest reduction in price (-60%), you need to have a sufficient amount of TFT in your wallet. The TFT is not locked and simply needs to be present in your wallet.
+
+
+For the capacity of a Presearch node, the amount is around 5000 TFT, covering 3 years that the workload could run.
+
+Note that a Presearch node requires about 3 days to stabilize.
+
+
+
+
+### Can I use the same seed phrase for my mainnet and testnest accounts? How can I transfer my TFT from mainnet to testnet or vice versa?
+
+Yes, you can use the same seed phrase for you main and testnet accounts. They will have the same address on each chain but they are really separate accounts. It is much like using the same wallet on Ethereum and BSC for example.
+
+To transfer your TFT from mainnet to testnet or vice versa, you need to send your TFT to the Stellar chain first. Let's say you want to transfer TFT from the mainnet to the testnet. Here are the steps:
+
+* Open your mainnet profile on the [mainnet Dashboard](https://dashboard.grid.tf/), on the left menu, choose Portal and then Swap. Click the button Withdraw. Send your TFTs from the mainnet address to your Stellar address.
+
+* Open your testnet profile on the [testnet Dashboard](https://dashboard.test.grid.tf/), on the left menu, choose Portal and then Swap. Click the button Deposit. Send your TFTs from your Stellar address to the testnet address. You can use the QR code option to make the transfer.
+
+To go from testnet to mainnet, simply use the URLs in the opposite order.
+
+> Note that the fees are of 1 TFT per transfer.
+
+
+
+### Do I need a full or micro virtual machine (VM) when I run QSFS, quantum safe file system, on the ThreeFold Grid?
+
+QSFS can be run on both a full virtual machine or a micro virtual machine (VM). The QSFS is a "mountable" object, like a disk. It's defined in its own block, then specified as a mount within a VM.
+
+
+## Linux, Github, Containers and More
+
+### Where should I start to learn more about Linux?
+
+While the [ThreeFold Manual](https://www.manual.grid.tf/) would be a good place to learn about ThreeFold and how to deploy on the TF Grid, to learn specifically about Linux, a good place to start is the [Linux website](https://www.linux.org/). There you will find [many tutorials](https://www.linux.org/forums/#linux-tutorials.122).
+
+A general advice to learn Linux, and computers in general, is to develop the skill of finding answers by following your natural curiosity: most of the questions have been asked before and answers can be found through search engines.
+
+Before doing any web search, you can use the resources already on hand in the terminal. For any command, you can try adding `-h` or `--help` for a brief description of what it does and to see some commonly used arguments. Typing `man` and then the command name will bring up a more detailed manual, assuming it exists and is installed (e.g. `man sudo`).
+
+Some ThreeFold users also point out that using different AI and LLM can be very helpful in the process of learning Linux and computers in general.
+
+
+### How can I clone a single branch of a repository on Github?
+
+You can clone a single branch of a repository with the following line:
+
+```
+git clone --single-branch --branch branch_name https://github.com/GITHUB_ACCOUNT/REPOSITORY_NAME
+```
+
+
+
+## Grace Period (Status Paused)
+
+### The status of my deployment is paused, in grace period, how can I resume the deployment?
+
+When your wallet is running out of TFT to pay for deployments, your deployments will be paused.
+To resume your deployments, simply fill up your wallet with more TFT.
+
+
+
+### Once I refund my TF wallet, how long does it take for the deployment to resume from grace period?
+
+It can take around one hour to change the status from "paused" to "ok" and thus for the deployment to resume.
+
+
+
+### Can I SSH into my deployments when they are in grace period (i.e. when their status is paused)?
+
+While in grace period, you might not be able to SSH into your deployment. Refund your wallet to resume deployments.
+
+
+
+### How long is the grace period (i.e. when the deployment status is paused)?
+
+The grace period is 2 weeks. During this period, you can refill your wallet to resume your deployment.
+
+
+
+## Terraform
+
+### Working with Terraform, I get the following error: failed to create contract: ContractIsNotUnique. Is there a fix to this issue?
+
+This error happens when a contract with the same data is already active. For example, two conflicting contracts are not in the same deployment. You can try to change the data on the main.tf file to make sure it's not a self-conflicting terraform deployment.
+
+
+
+### I am working with Terraform. What do I have to write in the file env.tfvars?
+
+This env.tfvars should look like the following, with the proper content within the quotes:
+
+> MNEMONICS = "write your seed phrase"
+>
+> NETWORK = "write the main network"
+>
+> SSH_KEY = "write your ssh-key"
+
+Note that this could change based on your specific Terraform deployment.
+
+
+
+### I am working with Terraform and I am using the example in Terraform Provider Grid. How can I use the example main.tf file with environment variables? Why am I getting the message Error: account not found, when deploying with Terraform?
+
+This Q&A is linked with the main.tf file in the [QSFS section of the ThreeFold Tech repository](https://github.com/threefoldtech/terraform-provider-grid/blob/development/examples/resources/qsfs/main.tf).
+
+This error happens when you did not properly set your environment variables.
+In the main.tf file, add those lines at the top:
+
+> variable "MNEMONICS" {
+>
+> type = string
+>
+> description = "The mnemonic phrase used to generate the seed for the node."
+>
+> }
+
+> variable "NETWORK" {
+>
+> type = string
+>
+> default = "main"
+>
+> description = "The network to connect the node to."
+>
+> }
+
+> variable "SSH_KEY" {
+>
+> type = string
+>
+> }
+
+Within the file main.tf, set those lines:
+
+> provider "grid" {
+>
+> mnemonics = "${var.MNEMONICS}"
+>
+> network = "${var.NETWORK}"
+>
+> }
+
+> env_vars = {
+>
+> SSH_KEY = "${var.SSH_KEY}"
+>
+> }
+
+Note: Make sure that you properly set your variables in the file env.tfvars.
+
+
+## Users Troubleshooting and Error Messages
+
+### When deploying a virtual machine (VM) on the ThreeFold Grid, I get the following message after trying a full system update and upgrade: GRUB failed to install to the following devices... Is there a fix to this issue?
+
+When deploying a virtual machine and doing a full system update and upgrade (apt update, apt upgrade), if you get the error: GRUB failed to install to the following devices /dev/vda15, try this to fix it:
+
+> apt-mark hold grub-efi-amd64-signed
+
+This should fix the issue.
+
+
+### While deploying on the TF Dashboard, I get the following error :"global workload with the same name exists: conflict". What can I do to fix this issue?
+
+This error happens if you deployed a workload on the TF Dashboard with the same deployment name of a previous deployment. In general, you can simply refresh the TF Dashboard browser page and/or change the deployment name for a new one. This should fix the issue.
+
+
+
+## ThreeFold Connect App
+
+### TF Connect App is now asking for a 4-digit password (PIN). I don't remember it as I usually use touch or face ID to unlock the app. What can I do?
+
+When you set up your the app, you are asked a 4-digit password (PIN). After some time, the app will be asking for this PIN when users may have been exclusively using touch/face ID. You can reset it by recovering the account with your seedphrase.
+
+
+### Is there a way to have more than one wallet in TF Connect App?
+
+Yes, this is perfectly possible. You can have multiple wallets in the TF Connect app. You can have multiple wallets for the Stellar network and multiple wallets for Polkadot Substrate.
+
+For example, you can create a wallet on the Stellar Blockchain and import it on TF Connect App with the function *Import Wallet*. Just copy the seedphrase and it will be imported in TF Connect App.
+
+Note: There will not be an automatic function in the app to create a new wallet. You must do it manually.
+
+
+
+### What is the difference between 10.x.y.z and 192.168.x.y addresses?
+
+The addresses 10.x.y.z and 192.168.x.y are on the private network.
+
+A 10.x.y.z address is just as valid as a 192.168.x.y address for a private network. If all the devices get a 10.x.y.z network, your private network should work properly. But if some of your devices are getting 10.x.y.z addresses while others are getting 192.168.x.y addresses, then you have two conflicting DHCP servers in your network.
+
+
+
+# DEVELOPERS FAQ
+
+## General Information for Developer
+
+### Does Zero-OS assign private IPv4 addresses to workloads?
+
+No. Zero-OS will request two IP address from the DHCP. If you only have one physical NIC connected, Zero-OS will assign the second IP address as a virtual device.
+
+
+
+### Why does each 3Node server have two IP addresses associated with it?
+
+Each node has two IP adresses.
+
+One is for for Zero-OS and one is for for the DMZ (demilitarized zone, sometimes referred to as a perimeter network or screened subnet). This separates public/private traffic from each other.
+
+Note: Zero-OS will request two IP address from the DHCP. If you only have one physical NIC connected, Zero-OS will assign the second IP address as a virtual device.
+
+### Can Zero-OS assign public IPv4 or IPv6 addresses to workloads?
+Yes it can provide both standard and Yggdrasil connections.
+
+
+### What does MAC mean when it comes to networking?
+MAC means *media access control*. It is a unique hardware ID. It helps the network to recognize your machine. It is then, for example, possible to assign a specific and fixed IP address to your hardware.
+
+
+
+
+### I am a developer looking for a way to automatically convert BSC tokens into TFT. Could you please share tips on how to swap regular tokens into TFT, on backend, without and browser extensions, via any platform API?
+
+TFT is implemented as a cross-chain asset (BToken) on BSC.
+
+Swapping via Pancakeswap directly without a browser and extension can be done by calling the Pancakeswap's BSC contract directly. Note that this requires some EVM knowledge on calling contracts.
+
+TFT is a standard ERC-20 contract, called BEP20 on BSC. You will thus need to call the approved method in the TFT contract to allow Pancakeswap to transfer TFTs from your account.
+
+The TFT contract allows you to approve a large amount and when Pancakeswap makes the transfer, the TFT will be deducted from the transaction. This will not reset so you will not have to make a call to approve every swap. This will thus save some gas.
+
+
+
+## Test Net
+
+### Can I get some free TFT to test on Test Net
+
+The TFT on Test Net is real TFT. There are ways to get free TFT to explore Test Net, such as joining the Beta Tester Group. More information [here](https://forum.threefold.io/t/join-the-grid-3-0-beta-testers-group/1194/21).
+
+
+
+# FARMERS FAQ
+
+## TFT Farming Basics
+
+### My Titan is v2.1 and the ThreeFold Grid is v3., what is the distinction?
+
+Titan v2.1 is the hardware. Before v2.1, there was 2.0. The hardware currently being shipped is the Titan v2.1.
+
+When you read v3, it refers to the ThreeFold Grid. Titans are now being sent ready for TF Grid v3 so they are being referred to as Titan v3. In short, the current Titans are v2.1 hardware ready for ThreeFold Grid v3.
+
+
+
+### When will I receive the farming rewards for my 3Nodes?
+
+Farming rewards are usually sent around the 8th of each month. This can vary slightly because the verification process is not yet fully automated.
+
+For more information on the minting process, read the next [QnA](#what-is-the-tft-minting-process-is-it-fully-automated).
+
+
+
+### What is the TFT minting process? Is it fully automated?
+
+Minting is based on blockchain data according to strict rules that are carried out by computers with humans involved only to check for errors and to sign the resulting transactions.
+
+There is a human verification mechanism through multisignatures for calculations done on the data as stored in the blockchain. This explains the timing differences when it comes to the monthly farming rewards distribution, since enough people need to sign off.
+
+The detailed minting process for V3 is as follow:
+
+- TFChain, ThreeFold's blockchain, has all the details about capacity provided by the nodes.
+- TFChain is used to track uptime.
+- Zero-OS reports to TFChain.
+- The code in [this repo](https://github.com/threefoldtech/minting_v3) uses the information from the blockchain to calculate the TFT to be minted.
+- A proof of what needs to be minted and why is created. This proof is then sent to our guardians.
+- The guardians need to double check the execution and the minting report. This is like a human check on the automated process.
+- The guardians need to sign. Only when consensus is achieved the minting as suggested will happen. This allows human to check the code.
+
+It is important to understant that TFChain tracks the capacity and uptime and is the source for the minting.
+
+Note: Additional auditing code will be added in V4 (i.e. special code generated at runtime for verification) using security primitives on motherboards.
+
+For more information on the minting periods, read this [QnA](#what-is-the-start-and-end-of-the-current-minting-period-what-are-the-minting-periods-for-threefold-farming-in-2023).
+
+
+
+### What should I do if I did not receive my farming rewards this month?
+
+If you did not receive your farming rewards, please contact us via our live chats. We will then investigate the situation.
+
+You can find our live chat option on the TF Connect App ([Android](https://play.google.com/store/apps/details?id=org.jimber.threebotlogin&hl=en_CA&gl=US), [iOS](https://apps.apple.com/us/app/threefold-connect/id1459845885)), the [TF Forum](http://forum.threefold.io/), [TF Dashboard](https://dashboard.grid.tf/) and the [TF website](https://threefold.io/).
+For the TF Connect app, select the Support option.
+For the other websites, it is a blue icon on the right bottom part of the page.
+
+The easiest way to contact the ThreeFold support is to use [this link](https://threefoldfaq.crisp.help/en/) and click on the chat icon.
+
+### What is the TFT entry price of my 3Node farming rewards?
+
+Currently, the TFT entry price of the 3Nodes' farming rewards is 8 cents (0.08 USD).
+
+
+
+### What is the necessary uptime for a 3Node per period of one month?
+
+Note that as of now, rewards are proportional to the uptime, so (e.g.) 40% uptime farms 40% of the total uptime period.
+
+When implemented : For certified Titans 3Nodes, it is 97% uptime per month (21.6h). For DIY 3Nodes, it is 95% uptime per month (36h). For professional certified 3 nodes, it is 99.5% uptime per month (3.6h).
+
+
+### How can I check the uptime of my 3Nodes? Is there a tool to check the uptime of 3Node servers on the ThreeFold Grid?
+
+You can go on the [ThreeFold Dashboard](https://dashboard.grid.tf/), and select "Farms" in the Portal menu. When you select a specific 3Node in your farm, you can see a visual graph of the 3Node's uptime of the last month. By clicking on "Node Statistics", you can see past and present uptime periods.
+
+
+
+### I set up a 3Node in the middle of the month, does it affect uptime requirements and rewards?
+
+You will still need to meet the required uptime (95% for DIY, etc.3), but the period starts when you connected the 3Node for the first time, instead of the usual start of the month. This only applies the first month. Afterwards, the uptime is calculated on the whole month. *Read last Q+A
+
+
+### What is the difference between a certified and a non-certified 3Node?
+
+A certified 3Node will receive 25% more reward compared to a non-certified 3Node.
+You can visit the [ThreeFold Marketplace](https://marketplace.3node.global/) for more information on certified 3Nodes.
+DIY certified 3Nodes are a future possibility.
+
+
+
+### What are the different certifications available for 3Node servers and farms? What are the Gold and Silver certifications?
+
+3Nodes can be certified. You can buy certified 3Nodes [here](https://marketplace.3node.global/).
+
+Farms can also be certified. The certifications are: [gold certified farming](https://forum.threefold.io/t/gep-gold-certified-farming-specs-closed/2925) and [silver certified farming](https://forum.threefold.io/t/silver-booster-request-for-comments/3416).
+
+Note that gold and silver certifications are still being discussed. Join the discussion on the [ThreeFold Forum](http://forum.threefold.io/).
+
+
+### What is the difference between V2 and V3 minting?
+
+V2 is being sunset. New miners should directly onboard to V3.
+On the tokenomics side, V2 rewards decrease as the difficulty level increases. For the V3 rewards, the rewards are constant for 5 years. In short, V3 is more profitable in the long run. For more information, read [this post](https://forum.threefold.io/t/comparison-v2-vs-v3-minting/2122).
+
+
+
+### What is the TFT minting address on Stellar Chain?
+
+The TFT minting address on Stellar Chain is the following: GBOVQKJYHXRR3DX6NOX2RRYFRCUMSADGDESTDNBDS6CDVLGVESRTAC47
+
+You can see it on the Stellar Explorer [here](https://stellar.expert/explorer/public/account/GBOVQKJYHXRR3DX6NOX2RRYFRCUMSADGDESTDNBDS6CDVLGVESRTAC47).
+
+
+
+### Can Titans and DIY 3Nodes share the same farm?
+
+Yes. It's one big ThreeFold family! A farm can have several 3Nodes (Titans or DIY) and each 3Node can be linked to only one farm.
+
+
+
+### Do I need one farm for each 3Node?
+
+No. You only need one farm. One farm can have multiple 3Nodes. When setting your farm, you will add an address for the farming rewards. All farming rewards from each 3Node of your farm will be sent to this address. Note that you can choose to have more than one farm. It is up to you.
+
+
+
+### Can a single farm be composed of many 3Nodes?
+
+Yes. You can have many 3Nodes on the same farm.
+
+
+
+### Can a single 3Node be on more than one farm?
+
+No, this is not possible.
+
+
+
+### Do I need one reward address per 3Node?
+
+You do not need more than one address linked to one farm. All of your 3Nodes can be connected to the same farm. Rewards per each 3Nodes will all be sent to the address linked to your farm.
+
+
+
+### How can I access the expert bootstrap mode for Zero-OS?
+
+You can access the expert bootstrap mode for Zero-OS at this link: [https://bootstrap.grid.tf/expert](https://bootstrap.grid.tf/expert).
+
+
+
+### When it comes to the Zero-OS bootstrap image, can I simply duplicate the first image I burnt when I build another 3Node?
+
+Yes. What is needed on this bootstrap image is to have the proper farm ID. The bootstrap image will be the same for all your different 3Nodes. It's a good TF farming practice to leave the bootstrap image plugged in the 3Node at all time.
+
+
+
+### If a node is unused for certain time (e.g. many months offline), will it be erased by the Grid?
+
+No, nodes only get deleted if the farm owner chooses to do so. Old "nodes" are really just entries in TF Chain and TF Chain does not modify or delete this data without external input.
+
+
+
+### Can a farm be erased from TF Grid?
+
+No, this is not possible. In the future, we will implement some features in order to allow the cleaning of unused farms. As of now, this is not possible. Also, an old farm does not take resources on the TF Grid, or very little.
+
+
+
+### On the ThreeFold Connect App, it says I need to migrate my Titan farm from V2 to V3. What do I have to do? How long does this take?
+
+To migrate, read [this documentation](https://forum.threefold.io/t/what-to-do-if-your-farm-is-still-on-grid-v2-closed/3761).
+
+
+
+
+### How can I migrate my DIY farm from V2 to V3?
+
+Create a new [bootstrap image](https://bootstrap.grid.tf/) using your new V3 Farm ID. To create a new V3 Farm ID, you can use the ThreeFold Connect App or the ThreeFold Dashboard.
+
+
+
+### What does the pricing policy ID of a farm represent?
+
+The pricing policy is the definition of how the network bills for workloads (pricing for each resource type, discounts, etc.).
+
+
+
+### What is the difference between TiB and TB? Why doesn't the TF Explorer shows the same storage space as my disk?
+
+Terabyte (TB) and Tebibyte (TiB) are units of digital information used to measure storage capacity and data transfer rate. While terabyte is a decimal standard unit, Tebibyte is binary.
+
+* 1 TB = 1000^4 bytes
+* 1 TiB = 1024^4 bytes.
+
+There are thus more bytes in 1 TiB than in 1 TB.
+1 TiB is equal to 1.099511627776 TB.
+
+You can play around these 2 notions by exploring this [TiB-TB converter](https://www.dataunitconverter.com/tebibyte-to-terabyte/).
+
+You can also check [this table](https://www.seagate.com/ca/en/support/kb/why-does-my-hard-drive-report-less-capacity-than-indicated-on-the-drives-label-172191en/) to compare different OS system's storage representation.
+
+
+
+## Farming Rewards and Related Notions
+
+### What are the rewards of farming? Can I get more rewards when my 3Node is being utilized?
+
+By connecting a 3Node to the Grid, you get Farming Rewards. If you set a public IP address for the Grid to use, you will receive additional rewards when your 3Node is being utilized by users on the Grid. All rewards are paid in TFT. To know the potential rewards, use the [simulator](https://simulator.grid.tf/). More information on sales channel will be communicated in the future.
+
+
+
+### How can I know the potential farming rewards for Grid Utilization?
+
+Go on the [ThreeFold simulator](https://simulator.grid.tf/), enter your 3Node resources, check the Public IP address. This will enable farming rewards from the parameter NU Required Per CU. Check the difference in the farming rewards per month. Note that you will need a Public IP address.
+
+
+
+### What is the easiest way to farm ThreeFold tokens (TFT)?
+
+Buy a [certified 3Node](https://marketplace.3node.global/index.php). This is more or less *plug n play*! You can also build a [DIY 3Node](#what-are-the-general-requirements-for-a-diy-3node-server). It's fun and there are many resources to help you along the way.
+
+
+
+### When do I receive my rewards?
+
+They are distributed once a month, around the 8th*. Distributions are not daily, or after a certain threshold. Note that upcoming minting rules may have a 24 month lockup or until 30% utilization for 3 months on your 3Node.
+
+*This can change slightly depending on the current situation.
+
+
+
+### Do farming rewards take into account the type of RAM, SSD, HDD and CPU of the 3Node server?
+
+No. The farming rewards do not take into account the specific type of RAM, SSD, HDD and CPU. The farming rewards take into account the quantity of storage and compute units (e.g. TB of SSD/HDD, GB of RAM, # of virtual cores).
+
+
+### Can I send my farming rewards directly to a crypto exchange?
+
+This is not possible. When you send tokens to a crypto exchange, you need to include a memo with your wallet address and the current farming rewards system of ThreeFold is already using that memo space to send the correct farming information.
+
+
+### Do I need collateral to farm ThreeFold tokens?
+
+Many decentralized data projects require collateral, but not ThreeFold. There is an ongoing discussion on collateral. Join the discussion [here](https://forum.threefold.io/t/should-tft-collateral-be-required-for-3nodes/3724).
+
+
+### Can I add external drives to the 3Nodes to increase rewards and resources available to the ThreeFold Grid?
+
+As of now, you cannot add external drives to a 3Nodes. It is not yet supported. It might be in the future and we will update the FAQ if/when this happens.
+
+
+
+### Do I have access to the TFT rewards I receive each month when farming?
+
+For now, V3 farming rewards are distributed as TFT on Stellar and they are immediately available. The lock system will be implemented on chain. Tokens will be staked to your address until unlock conditions are met. Conditions are: 2 years of farming or 30% of proof-of-utilization for 3 months per 3Node.
+
+
+
+### What is TFTA? Is it still used?
+
+Note that on V3, TFTA will not be issued anymore.
+
+
+
+### Is there a way to certify a DIY 3Node? How can I become a 3Node certified vendor and builder?
+
+As of now, only certified ThreeFold partners can certify a 3Node. You could become a ThreeFolder vendor and offer certified 3Node. Read more [here](https://marketplace.3node.global/index.php?dispatch=companies.apply_for_vendor).
+
+
+
+### Does it make sense to make my farm a company?
+
+There is no general answer to this. Here's what a ThreeFold member thinks about this if you are living in the USA. Check for your current location if this makes sense for you. DYOR.
+
+> "Definitely do this project as a business entity. You can write off equipment, utilities and a portion of your home's square footage if your are hosting the equipment at home." TFarmer
+
+
+
+### What is the difference between uptime and downtime, and between online and offline, when it comes to 3Nodes?
+
+Uptime and status are two different things. As long as the 3Node is powered on, its uptime does not reset. Its status changes to offline if it hasn't made an uptime report in the last two hours. Even if after more than two hours, the 3Node isn't yet online, the uptime does not reset. Uptime is a function of the system being powered on.
+
+
+### My 3Node server grid utilization is low, is it normal?
+
+This is normal. Currently, not all 3Nodes are being utilized by users on the ThreeFold Grid. Each month, more and more utilization happens and the ThreeFold Grid continually expans. To read on the ThreeFold Grid utilization and expansion, check this [TF Forum post](https://forum.threefold.io/t/grid-stats-new-nodes-utilization-overview/3291).
+
+
+
+
+## 3Node Farming Requirements
+
+
+### Can I host more than one 3Node server at my house?
+
+Yes, but do not host more than your bandwidth can support.
+
+
+
+### Is Wifi supported? Can I farm via Wifi instead of an Ethernet cable?
+
+No. Wifi is not supported by Zero-OS due to a number of issues, like reliability, performance, configuration requirements and driver support. It's all about Ethernet cables here.
+
+
+
+### I have 2 routers with each a different Internet service provider. I disconnected the ethernet cable from one router and connected it to the other router. Do I need to reboot the 3Node?
+
+You do not need to reboot. The 3Node will be able to reconnect to the ThreeFold Grid.
+
+
+
+
+### Do I need any specific port configuration when booting a 3Node?
+
+No, as long as the 3Node is connected to the Internet via an ethernet cable (wifi is not supported), Zero-OS will be able to boot. Usually with the DHCP, it automatically assigns an IP address.
+
+
+
+### How much electricity does a 3Node use?
+
+A small DIY 3Node based on a compact office computer will draw under 20W. A full size server will draw around 100W idling. Note that a 3Node actively used on the Grid (proof-of-utilization) will draw more power, but also generate passive income on top of farming if you have a public IP address.
+
+For more information, read thes section [Calculate the Total Electricity Cost of Your Farm](../farmers/farming_optimization/farming_costs.md#calculate-the-total-electricity-cost-of-your-farm) of the Farming Guide.
+
+
+
+### Has anyone run stress tests to know the power consumption at heavy load of certain 3Nodes?
+
+The community is starting to gather some data on this. As of now, we know that a R720 with 2x2690v2 cpu, 4TB NVME SSE P4510 and 320GB ram will draw 390W @100% load. With 2x2650L v2, it's around 300W with fans at full speed. More info will be added as we gather more data.
+
+### Can the Titan 3Node be run on PoE? (Power Over Ethernet)
+
+Titans don't come equipped for Power Over Ethernet (PoE). If you have a NUC based Titan there are some PoE lids that might be compatible.
+
+
+
+### What is the relationship between the 3Node's resources and bandwidth?
+
+A 3Node connects to the ThreeFold Grid and transfers information, whether it is in the form of compute, storage or network units (CU, SU, NU respectively). The more resources your 3Nodes offer to the Grid, the more bandwidth will be needed to transfer the additional information.
+
+
+### What is the bandwidth needed when it comes to running 3Nodes on the Grid?
+
+The strict minimum for one Titan is 1 mbps of bandwidth.
+
+If you want to expand your ThreeFold farm, you should check the following to make sure your bandwidth will be sufficient when there will be Grid utilization.
+
+> min Bandwidth per 3Node (mbps) = 10 * max((Total SSD TB / 1 TB),(Total Threads / 8 Threads),(Total GB / 64 GB)) + 10 * (Total HDD TB / 2)
+
+This equation means that for each TB of HDD you need 5 mbps of bandwidth, and for each TB of SSD, 8 Threads and 64GB of RAM (whichever is higher), you need 10 mbps of bandwidth.
+
+This means a proper bandwidth for a Titan would be 10 mbps. As stated, 1 mbps is the strict minimum for one Titan.
+
+The bandwidth needed for a given 3Node is not yet set in stone and you are welcome to participate in ongoing the [discussion on this subject](https://forum.threefold.io/t/storage-bandwidth-ratio/1389) on the Forum.
+
+
+
+### Can I run Zero-OS on a virtual machine?
+
+You can. But you won't farm TFT. To farm TFT, Zero-OS needs to be on bare metal.
+
+
+### Is it possible to build a DIY 3Node with VMWare VM ?
+
+It wouldn't be possible to get farming rewards from such configuration. You need to run a 3Node Zero-OS on bare metal and no virtual machine is permitted. Indeed, to farm TFT you need bare metal. Virtual Machine will not work. Furthermore, all disks of a 3Node need to be wiped completely, thus no other OS can be on the 3Node.
+
+It would be possible to simply set Zero-OS as a VM on VMWare VM, but it wouldn't farm as stated.
+
+
+### Can I run a 3Node on another operating system, like Windows, MAC or Linux?
+
+No. ThreeFold runs its own operating system (OS), which is Zero-OS. You thus need to start with completely wiped disks.
+
+### What is the minimum SSD requirement for a 3Node server to farm ThreeFold tokens (TFT)?
+
+You need a theoretical minimum of 500 GB SSD on a desktop or server. Less could work.
+
+
+
+### Is it possible to have a 3Node server running on only HDD disks?
+
+This is not possible. A 3Node needs at least one SSD disk of 500 GB.
+
+
+
+## Building a 3Node - Steps and Details
+
+
+### How can I be sure that I properly wiped my disks?
+
+A wiped disk has:
+- no label
+- no partition
+- no filesystem
+- only zeroes
+
+On Linux to see if the disk is only composed of zeroes, use the command line (example with disk sda):
+
+> cmp /dev/sda /dev/zero
+
+If there is only zeroes, you should get the output:
+
+> cmp: EOF on /dev/sda
+
+You can also use the command line:
+
+> sudo pv /dev/sda | od | head
+
+Here are some useful command lines for Linux to make sure there are no partitions, no labels, no filesystems and that the disks are filled with zeroes only:
+
+> sudo fdisk -l
+>
+> sudo fdisk -lf
+>
+> sudo parted -l
+>
+> sudo parted /dev/sda 'print’
+>
+> lsblk --f
+>
+> lsblk --raw
+
+
+
+### If I wipe my disk to create a new node ID, will I lose my farming rewards during the month?
+
+No, you wouldn't lose any farming rewards. You will get both the rewards for your previous node ID's uptime and also the new node ID's uptime.
+
+Note that this is the case with the current farming rewards based on total uptime, without any minimum threshold.
+
+
+
+### My disks have issues with Zero-OS and my 3Nodes. How can I do a factory reset of the disks?
+
+> Warning: this is destructive. It erases the disk sda in this example.
+
+Boot a Linux in Try mode and run the following command:
+
+> sudo badblocks -svw -b 512 -t 0x00 /dev/sda
+
+*In this example, the disk selected is sda. Choose the proper disk name in your current situation, e.g. sdb, sdc..).
+
+This line will read and (over)write zeroes (0x00) everywhere on the disk.
+
+To understand the line of code, note that 512 is the block size and that without -b BLOCKSIZE, the process would simply go slower.
+0x00 represents the zero byte.
+
+
+
+This will take some time, but it should reset the disk and hopefully fix any issues.
+
+Note: it takes the 3Node server around 1 percentage per minute for a 2TB SSD disk to accomplish the badblock operation.
+
+
+
+### Before doing a bootstrap image, I need to format my USB key. How can I format my USB key?
+
+*Note that BalenaEtcher will format and burn your bootstrap image in the same process. See next Q+A for more details.
+
+Windows: This is done easily with diskpart. Here's all the coding needed (with disk X as an example, make sure you choose the correct disk): run Command Prompt as an administrator (right-click option), write *diskpart*, then in diskpart write *list disk*, choose your disk and write *select disk X*, write *clean*, write *create partition primary*, write *format fs=fat32*, then write *assign*. If for any reason, the process doesn't work, quit diskpart and redo the whole thing. This should fix any bug. Cautious with diskpart, it's destructive.
+
+MAC: This is done easily with Disk Utility. Go in Disk Utility. Select your USB key, click on *erase* on the top, write a name for your USB key, then choose a format (MS-DOS (FAT) if USB key < 32GB, exFAT if USB key > 32GB), then click *erase*, then click *done*.
+
+LINUX: In the Terminal, write *df*, find your disk (here we use sdX), write *sudo umount /dev/sdX*, write this line (with the proper FORMAT) *sudo mkfs.FORMAT /dev/sdX* [FORMAT= vfat for FAT32, ntfs for NTFS, exfat for exFAT], then finally verify the formatting by writing *sudo fsck /dev/sdX*.
+
+### What do you use to burn (or to load) the Zero-OS bootstrap image onto a USB stick?
+
+For MAC, Linux and Windows, you can use [BalenaEtcher](https://www.balena.io/etcher/) to load/flash the image on a USB stick. This program also formats the USB in the process. Rufus can also be used for Windows.
+
+Also, got Linux systems, you can transfer the dowloaded image with the dd command: *dd if=created_boot_loader_file.img of=/dev/sd?* where the input file is the downloaded file from http:/bootloader.grid.tf and the output file (device) is the USB stick device.
+
+
+### Should I do a UEFI image or a BIOS image to bootstrap Zero-OS?
+
+It depends on your 3Node's system. Newer computers and servers will accept UEFI. If it does not work with UEFI, please try with the options ISO (BIOS CD/DVD) or USB (BIOS image) on the [ThreeFold bootstrap website](https://bootstrap.grid.tf). Read the next Q+A for more information on BIOS/UEFI.
+
+
+
+### How do I set the BIOS or UEFI of my 3Node?
+
+You can read this [documentation](../farmers/3node_building/5_set_bios_uefi.md) to learn more about BIOS and UEFI settings for a DIY 3Node.
+
+
+
+### For my 3Node server, do I need to enable virtualization in BIOS or UEFI?
+
+Yes, you should enable virtualization. On Intel, it is denoted as *CPU virtualization* and on ASUS, it is denoted as *SVM*. Make sure virtualization is enabled and look for the precise terms in your specific BIOS/UEFI.
+
+
+
+### How can I boot a 3Node server with a Zero-OS bootstrap image?
+
+Plug the USB key containing the Zero-OS bootstrap image with your farm ID then power on your 3Node. If the BIOS/UEFI is set correctly and the disks are all wiped, it should boot correctly the first time. If you have any problem booting your 3Node, read the section [Troubleshooting and Error Messages](#troubleshooting-and-error-messages) of the FAQ.
+
+
+
+### The first time I booted my 3Node server, it says that the node is not registered yet. What can I do?
+
+The first time you boot a 3Node, it will be written: “This node is not registered (farmer *: NameOfFarm). This is normal. The Grid will create a node ID and you will be able to see it on screen. This can take a couple of minutes.
+
+If after some time (couple hours), the 3Node doesn't get registered, there might be something off with the Grid connection. You can then try to reboot the 3Node, or wait and boot it later. If it persists, you can check the rest of the Troubleshooting section of the Farmer FAQ, or ask around the ThreeFold Telegram Farmer chat or the ThreeFold chat for help.
+
+
+
+### The first time I boot my 3Node, the node gets registered but it says cache disk : no ssd. What can I do?
+
+This probably means that you either haven't connected a SSD or that you need to wipe your SSD disk(s). Zero-OS runs on bare metal and needs a minimum of one SSD disk (min 500GB & 50 GB per CU). You will see "cache disk : OK" when it works.
+
+
+
+### The first time I boot my 3 node, the node gets registered and it says cache disk : OK, but the table System Used Capacity is empty. What can I do?
+
+Most of the time, just wait and data will appear. If you want to be sure your 3Node is online on the Grid, you can check the [Node Finder](https://dashboard.grid.tf/), which fetch information every 2 hours. If it persist, first try to simply reboot your 3Node.
+
+
+
+### I have a relatively old server (e.g. Dell R710 or R620, Z840). I have trouble booting Zero-OS. What could I do?
+
+Sometimes, Zero-OS will not boot in UEFI mode on older servers. In that case, try to boot in BIOS mode. Use either a USB key or the CD/DVD optical drive (the 4th and 5th option on https://bootstrap.grid.tf/) and make sure to select BIOS and not UEFI mode in your server settings.
+
+
+### I connected a SATA SSD to a CD-DVD optical drive adaptor. My system does not recognize the disk. What can I do?
+
+Try to set AHCI mode instead of Legacy mode in SATA settings in the BIOS.
+
+
+### Can someone explain what should I put in the Public IP part of my farm? Should I just insert my Public IP and Gateway (given by my ISP)?
+
+Assuming you are a DIY farmer and operate from your home, this field can be left blank. You do not have to fill in any details.
+
+The add IP option is for farmers that have a block of IP addresses routed to their router (in data centers mostly) and want to present “dedicated IP” addresses for deployments. For more information on how to set the public configuration, go to [this link](https://library.threefold.me/info/manual/#/manual__public_config).
+
+
+
+
+
+## Farming Optimization
+
+### What is the difference between a ThreeFold 3Node and a ThreeFold farm? What is the difference between the farm ID and the node ID?
+
+A farm is a composition of one or many 3Nodes. A 3Node is a computer connected to the ThreeFold Grid. Each farm has its farm ID and each 3Node has its node ID.
+
+
+
+### How can I know how many GB of SSD and RAM do I need?
+
+You need 50 GB of SSD per compute units (CU) and a minimum of 500 GB SSD and 2 GB of RAM per 3Node.
+
+A 3Node has, in general, 2 compute units (CU) per thread. Thus, for peak optimisation, you need 100 GB SSD and 8GB RAM per thread.
+
+### What is the optimal ratio of virtual cores (vcores or threads), SSD storage and RAM memory? What is the best optimization scenario for a 3Node, in terms of ThreeFold tokens (TFT) farming rewards?
+
+In short, for peak optimization, aim for 100 GB SSD of storage and 8GB RAM of memory per virtual core (vcore or thread).
+
+For example, a 32 threads (32 vcores) 3Nodes would need 3.2 TB SSD and 256GB RAM to be optimal, reward-wise.
+That is: 32 * 100 = 3200 GB SSD = 3.2TB SSD, and 32 * 8 = 256 GB RAM total.
+
+Adding more GB of RAM would not increase your TFT rewards. You would need more vcores if you want to expand.
+
+NB: This is purely based on reward considerations. Some users might need different ratios for different specific uses of the Grid.
+
+
+
+### What does TBW mean? What is a good TBW level for a SSD disk?
+
+TBW means Terabytes Written. TBW directly measures how much you can write cumulatively into the drive over its lifetime. For your 3Node, it can be a good idea to prioritize a minimum ratio of 500 TBW per 1TB for SSD.
+
+*Note that TBW is not a technical specification, but a general claim from the manufacturer. For this reason, it can also be good to check the warranty of the disk. For example, if a manufacturer offers a 5-year warranty to its product, it indicates that the company thinks its product will last a long time.
+
+
+### Are SATA and SAS drives interchangeable?
+
+This goes only one way. You can put a SATA drive in a SAS slot, but you can’t put a SAS drive in a SATA slot. See the [next question](#what-is-the-speed-difference-between-sas-and-sata-disks) for more information.
+
+
+### What is the speed difference between SAS and SATA disks?
+
+One of the big differences between SATA and SAS is the transfer speed. Using SATA disks with SAS cables, you will be limited by the SATA transfer speed.
+
+* Sata I : 150 MB/s
+* Sata II : 300 MB/s
+* Sata III : 600 MB/s
+* SAS : 600-1500 MB/s
+
+Note: You will most probably need to re-flash the raid card if you use the front panel disks (onboard storage) of your server.
+
+
+
+
+### Is it possible to do a graceful shutdown to a 3Node server? How can you shutdown or power off a 3Node server?
+
+There are no "graceful" shutdowns of 3Nodes. You can shutdown a 3Node from the software side. You need to shut it down manually directly on the hardware. 3Nodes are self-healing and if they suddenly power down, no data or information will be lost.
+
+
+
+
+### Is it possible to have direct access to Zero-OS's core to force a reboot?
+
+No, this is not possible. The general philosophy with Zero-OS is: no shell, no GUI, and no remote control. In other words, anything that could potentially provide attack surface is off the table. This ensures a high security level to Zero-OS and the ThreeFold Grid in general. To reboot a 3Node, you have to do it manually.
+
+
+
+### Do I need some port forwarding in my router for each 3Node server?
+
+No, this is not needed.
+
+
+### Are there ways to reduce 3Node servers' noises?
+
+To reduce the noise, you can remove all the unnecessary cards in the servers as well as the HDD disks if you don't use them. Unplugging the SAS cables can also help. You can also set the fans to adjust their speed instead of being constant.
+
+At the end of the day, servers were manufactured for durability and efficiency, and not for being quiet. Most servers are placed in server rooms where noise doesn't matter much.
+
+
+
+### I built a 3Node out of old hardware. Is it possible that my BIOS or UEFI has improper time and date set as factory default?
+
+Yes. Make sure you have the correct time and date in BIOS to avoid errors when trying to boot Zero-OS. It might not cause any problems, but sometimes it does.
+
+
+
+### I have rack servers in my ThreeFold farm. Can I set rack servers vertically?
+
+In general, it is not recommended to set rack servers vertically as they were designed to be laid flat in racks. That being said, if you want to set your rack vertically, here are some general rules to follow. Do so at your own risk.
+
+First, make sure the parts in the servers are well installed and that they will not fall if laid vertically. Second, and foremost, you want to make sure that there will not be any overheating. This means to make sure you don't block the front and rear of the unit, so heat can dissipate thought the vents.
+
+If you want to put the rack vertically with the longest side of the rack laying upward, having the power supply units (PSUs) on the very top will ensure that heat dissipate well.
+
+
+
+
+## Farming and Maintenance
+
+### How can I check if there is utilization on my 3Nodes?
+
+To see if there is utilization on your 3Node, you can consult the [TF Dashboard](https://dashboard.grid.tf/), go to the Farm section and consult the information under your 3Nodes. Note that the quickest way is to check if there are CPUs reserved.
+
+
+
+### Do I need the Zero-OS bootstrap image drive (USB or CD-DVD) when I reboot, or can I boot Zero-OS from the 3Node main hard drive?
+
+It is advised to keep the bootstrap image plugged in your 3Node. Once your node has been booted with Zero-OS via the USB key, you can remove the USB key, but if something happens and the node needs to reconnect with the network, it won’t be able to do so. We advise people to let the USB key always in so the node can reconnect with the network if needed.
+
+You need the bootstrap image device plugged in every time you reboot a 3Node and it's a good practice to keep it plugged in all the time. The technical explanation is: (1) at first boot, it creates a minimum requirement on SSD which is used as cache (2) each time the system restarts it reuses this SSD piece but Zero-OS bootstrap is also needed to download the last image. Indeed, image is not stored on the machine and also no boot loader is installed.
+
+
+### It's written that my node is using 100% of HRU. What does it mean?
+
+HRU stands for your HDD space available. It means that you are using 100% of the HDD space available, or equivalently that you have no HDD on your system.
+
+
+### On the ThreeFold Node Finder, I only see half of the virtual cores or threads my 3Node has, what can I do?
+
+Check in the BIOS settings and make sure you have enabled Virtual Cores (or Hyper Threading/Logical Cores).
+
+
+
+### Why are the 3Nodes' resources different on the ThreeFold Node Finder and the ThreeFold Dashboard?
+
+There is a difference because one shows the resources in GiB and the other in GB. It's just a way to display information, the resources are ultimately the same, but shown differently. 1 GiB = 1.073741824 GB.
+
+
+
+### How can I test the health of my disks?
+
+There are many ways to test the health of your disks. For some information on this, you can have a look at this [TF forum post](https://forum.threefold.io/t/testing-ssd-health/3436).
+
+
+
+### How can I transfer my 3Node from one farm to another?
+
+To transfer your 3Node from one farm to another, simply make a new bootstrap image of the new farm, connect the new bootstrap image to the 3Node and restart the node.
+
+Note that you can use [balenaEtcher](https://etcher.balena.io/) to format and burn the bootstrap image at the same time.
+
+
+
+### What do CRU, MRU, HRU and SRU mean on the ThreeFold Node Finder?
+
+CRU means the number of virtual cores. MRU means the GB of ram (memory). HRU means the HDD capacity storage and SRU means the SSD capacity storage.
+
+
+
+### I have more than one ThreeFold 3Node farm, but I want all my 3Nodes on only one farm. How can I put all my 3Nodes on one farm? How can I change the farm ID of my 3Node?
+
+If you have more than one 3Node associated with more than one farm, it is possible to put all the 3Nodes under only one farm.
+The following shows how to change the farm ID of your 3Node.
+
+For each 3Node, you need to create a new USB key Zero-OS boostrap image associated with the chosen farm ID. This will ensure your 3Node links to the chosen farm.
+To create a new USB key Zero-OS bootstrap image, you can download the bootstrap image [here](https://v3.bootstrap.grid.tf/) or you can clone an existing USB key bootstrap image associated with the chosen farm ID. In both case, you can use a software like [BalenaEtcher](https://www.balena.io/etcher/).
+Once you have the new USB key bootstrap image, plug it into the 3Node and reboot the server.
+After reboot, the 3Node will have the same node ID as before, but it will now be associated with the chosen farm ID.
+
+Note: If you want a new node ID for your 3Node, you will need to wipe the main SSD of your 3Node.
+
+
+
+### How can I know if my 3Node is online on the Grid?
+
+There are multiple answers to this.
+
+(1) You can plug a monitor to your 3Node and check directly its status.
+
+(2) You can check your 3Node status by using the Node Finder of the [ThreeFold Dashboard](https://dashboard.grid.tf/).
+
+(3) You can also use the unofficial, but highly useful [ThreeFold Node Status Bot](https://t.me/tfnodestatusbot) on Telegram.
+
+
+
+### I booted my 3Node and the monitor says it's online and connected to the Grid. But the ThreeFold Node Finder says it is offline? What can I do?
+
+The Node Finder refetches information every 2 hours. You can wait 2 hours and see if the problem persists. Then/or, you could reboot the 3Node and see if the Node Finder sees it as online.
+
+
+
+### My 3Node does show on the ThreeFold Node Finder, but not on the ThreeFold Dashboard, what can I do?
+
+If you're 3Node is correctly registered on the ThreeFold Grid but you cannot see it on the ThreeFold Dashboard, there can be many different ways to solve this issue.
+
+One way to fix this is to go in the TF Dashboard, select *change the address* and then simply re-paste the same address, then the extension will ask you to resign. Usually this fixes the issue.
+
+If the first method did not work, you can try to remove the account and add it back up on the Polkadot.js extension. Before doing so, make sure you have a back up of your seed phrase as you will need it to re-enter the account. Your 3Node should then appear.
+
+
+### If I upgrade my 3Node, will it increase my rewards?
+
+Yes. Use the simulator to verify the additional rewards. But note that, currently, upgrades are not recognized until the next minting cycle.
+
+
+### I booted my 3Node for the first time at the beginning of the month, then I did some upgrade or downgrade, will the ThreeFold Grid recognize the new hardware? Will it still be the same 3Node ID?
+
+Downgrades are counted in the current minting cycle. So you mint for the entire cycle at the downgraded specs.
+
+Minting only considers a single configuration per node per cycle, and that is the minimum configuration seen at any point during the cycle. With this logic in mind, upgrades will only be recognized at the next minting cycle. Note that your 3Node will have a new ID and new price entry if you changed the SSD containing the 3Node ID.
+
+
+
+
+
+
+### Is it possible to ask the 3Node to refetch the node information on the monitor?
+
+To refetch information, press " q " on the keyboard.
+
+
+
+### When does Zero-OS detect the capacity of a 3Node?
+
+Zero-OS only detects capacity right after it boots.
+
+
+
+
+### Where is the 3Node ID stored?
+
+For the current Zero-OS version, the node ID is stored in the first SSD you install on your 3Node*. If you change or erase this disk, this disk will lose its current 3Node ID.
+
+*Note that if, at the first boot, you put a SSD SATA and a SSD NVME at the same time, the node ID will be registered on the SSD SATA.
+
+
+
+
+### Is there a way to backup my node ID in order to restore a 3Node if the disk with the node ID gets corrupted or breaks down?
+
+Yes, you can do a backup, but as of now this process must be done manually.
+
+One way is to boot a Linux USB image in *Try* mode, open up the *File* folder of your disk that contains the node ID. Click on *+ Other Locations.* Then, open the folder that contains the folder *zos-cache* and open the folder *identityd.* In this folder, select the file *seed.txt* and make a copy of it in a safe place (USB key, notebook, e-mail, etc.). If the disk which contains your node ID is damaged, simply reboot the 3Node with a new disk and the Zero-OS bootstrap image. Your 3Node will connect to the Grid and assign a new node ID. Once this is done, reboot the 3Node, but this time with the Linux USB image, go in the same folder as stated before and replace the new *seed.txt* file with the old file. Reboot your 3Node with the Zero-OS bootstrap and you're done.
+
+
+
+### If I upgrade my 3Node, does it change the node ID?
+
+Upgrades won't change the node ID, unless you replace the SSD where the node ID is stored (see above for more info on this).
+
+
+
+
+### Does it make sense to recreate my node when the price drops?
+
+Short answer: no. Long answer: [click here](https://forum.threefold.io/t/does-it-make-sense-to-recreate-my-node-when-the-price-drops/).
+
+
+### My 3Node lost power momentarily and I had to power it back on manually. Is there a better way to proceed?
+
+In your BIOS, go in *Security Settings* and choose *Last* for *AC Power Recovery*. If you want, set a delay between 60 and 240 seconds. This will ensure your 3Node does not power on and off frantically if your power flickers on and off, thus potentially damaging the unit. On other BIOS, it's *After Power Loss*, and you should choose *Previous State*.
+
+*Depending on your 3Node, the parameter might have a different name.
+
+
+### Do I need to change the battery BIOS?
+
+It can be a good thing to change it when you buy an old desktop or server to make sure it lasts long. When the battery goes out of power, the 3Node won't have access to the BIOS settings if it loses power momentarily.
+
+
+
+### Do I need to enable UEFI Network Stack?
+
+You don't need to if you use a removable media (e.g. USB key) as a booting image. It is needed only if you boot from a PXE server on your network. You should keep this feature disabled. Enable it only if you know 100% what you are doing. Otherwise it might bring vulnerabilities in terms of network security.
+
+
+### I want redundancy of power for my 3 nodes. I have two PSU on my Dell server. What can I do?
+
+Make sure you enable the Hot Spare feature. This feature is accessible in iDRAC Settings - Power Configuration. Other servers might have this function, with a different name and configuration. Check the server's manual for more details.
+
+
+
+### Why isn't there support for RAID? Does Zero-OS work with RAID?
+
+RAID is a technology that has brought resilience and security to the IT industry. But it has some limitations that we at ThreeFold did not want to get stuck in. We developed a different (and more efficient way to store data reliably. Please have a look [here](https://library.threefold.me/info/threefold#/cloud/threefold__cloud_products?id=storage-quantum-safe-filesystem).
+
+This Quantum Safe Storage overcomes some of the shortfalls of RAID and is able to work over multiple nodes geographically spread on the TF Grid.
+
+
+### Is there a way to bypass RAID in order for Zero-OS to have bare metals on the system? (No RAID controller in between storage and the Grid.)
+
+Yes it is possible. "You can use the on board storage on a server without RAID. You can [re-flash](https://fohdeesha.com/docs/perc.html) the RAID card, turn on HBA/non-RAID mode, or install a different card. No need for RAID." @FLnelson It's usually easy to set servers such as a HP Proliant with the HBA mode. For Dell servers, you can either cross-flash the RAID controller with an “IT-mode-Firmware” (see this [video](https://www.youtube.com/watch?v=h5nb09VksYw)) or get a DELL H310-controller (which has the non-RAID option). Otherwise, you can install a NVME SSD with a PCIe adaptor, and turn off the RAID controller.
+
+
+
+### I have a 3Node rack server. Is it possible to use a M.2 to SATA adapter in order to put the M.2 SATA disk in the HDD bay (onboard storage)?
+
+Yes, it is possible. You will most probably need to bypass the RAID controller for Zero-OS to access properly the onboard storage. See previous question.
+
+
+
+### My 3Node uses only PCIe adapters and SSD NVME disks. Do I need the RAID controller on?
+
+The onboard RAID controller is not linked to your PCIe SSDs. In this case, you can switch the RAID controller off.
+
+
+
+### Can I change the name of my farm on polkadot.js?
+
+It’s possible to rename farms through the Polkadot UI. For mainnet, use [this link](https://polkadot.js.org/apps/?rpc=wss%3A%2F%2Ftfchain.grid.tf#/extrinsics).
+
+1. Under *using the selected account*, select the account that owns the farm
+2. Choose *tfgridModule* from the dropdown menu *submit the following extrinsic*,
+3. Select *updateFarm(id, name, pricingPolicyId)*
+4. Under *name: Bytes*, write the new farm name
+5. Finally, click on the bottom *Submit Transaction* at the bottom right of the screen
+
+
+
+### How can I delete a farm on polkadot.js?
+
+The ability to delete a farm was removed from TF Chain due to concerns that nodes could be left without a farm and thus cause problems with billing.
+
+
+
+### I try to delete a node on the TF Dashboard, but it doesn’t work. Is there any other way to proceed that could work?
+
+It’s possible to delete nodes through the Polkadot UI. For mainnet, use [this link](https://polkadot.js.org/apps/?rpc=wss%3A%2F%2Ftfchain.grid.tf#/extrinsics).
+
+1. Under *using the selected account*, select the account that owns the farm,
+2. Choose *tfgridModule* from the dropdown menu *submit the following extrinsic*
+3. Select *deleteNodeFarm(nodeId)*
+4. Under *id: u32*, write the ID of the 3Node you want to delete
+5. Finally, click on the bottom *Submit Transaction* at the bottom right of the screen
+
+
+
+### My 3Node has 2 ethernet ports in the back, with one written AMT above, what does it mean? Can I use this port to connect my 3Node to the ThreeFold Grid?
+
+First, let's define the term AMT. It means: Active Management Technology. Without going into too much details, it is to remotely access servers via the network at the BIOS level. Thus, you should plug the ethernet cable in the port next to AMT, and not into the AMT port. You can explore AMT properties if you want remote access to your server.
+
+
+
+### My 3Node is based on a the hardware Z600, Z620 or Z820, can I run it headless or without a GPU?
+
+For the Z600, there is a great [video on youtube](https://www.youtube.com/watch?app=desktop&v=JgBYbaT-N-w).
+
+For the Z620 and the Z820, you need to do some variation on the video above. In the BIOS, go in File, then Replicated Setup, and select Save to Removable Storage Device. This will save a text file on your USB key. Then, in the text file, go to Headless Mode and remove the * in front of Disable and put it in front of Enable. Save the file and then go back into BIOS. Now go in File, then Replicated Setup, and select Restore from Removable Storage Device.
+
+Running your 3Node without the GPU can save some power consumption as well as giving you one more extra slot for other hardware.
+
+
+### Is it possible to add high-level GPU on rack servers to farm more TFT?
+
+Some farmers had success installing GPUs such as the RTX3080 in servers as small as 2U (such as R730). Connections such as 250W 8-pin plug are needed on each riser. Generally, tower servers have more space to add high-level GPU.
+
+Note: GPU farming will be implemented in the future.
+
+
+
+### If I change farm, will my node IDs change on my 3Node servers?
+
+You can move 3Nodes between farms without losing the node IDs. You simply need to change the boot media with the new farm ID and then reboot the 3Node. The 3Node will have the same node ID but it will now be associated with the new farm ID.
+
+
+## Troubleshooting and Error Messages
+
+### Is it possible to access the Error Screen or Log Screen?
+
+Yes! On the Zero-OS console, hit alt-F2 to open up the Error/Log Screen, and hit alt-F3 to go back to the main screen.
+
+
+### What does it mean when I see, during the 3Node boot, the message: error = context deadline exceeded?
+
+In general, this message means that the ThreeFold Grid asked something to your 3Node, and your 3Node could not respond fast enough. It is usually necessary to read the following error message to understand the situation more specifically.
+
+
+### I try to boot a 3Node, but I get the error: "No Route to Host on Linux". What does it mean?
+
+There are many potential answers to this. Perhaps the Host is offline, the service isn't running. This is usually the reason with TF Grid. It means the Grid is not responsive. In this case, try to boot the 3Node later. If it persists ask TF Support.
+
+There can also be other reasons. You might have connected to the wrong port. Perhaps you have configured iptables to block connections on that port. Your DNS might be improperly configured. You might have an Incorrect Network or Host Configuration. Many troubleshoots are possible. Here's a [good place to start](https://www.maketecheasier.com/fix-no-route-to-host-error-linux/).
+
+
+
+### How can I fix the error: "Network configuration succeed but Zero-OS kernel could not be downloaded" when booting a 3Node?
+
+To fix the error "Network configuration succeed but Zero-OS kernel could not be downloaded", you can try to restart the router and reboot the 3Node. This usually fixes the issue. If this doesn't work, check if the router is still functional. The cause of this issue might be that your router is broken.
+
+
+
+### Using SAS disks, I get the error; "No ssd found, failed to register". What can I do to fix this?
+
+First make sure to wipe the disks and then boot your 3Node. If you've wiped the disks and it doesn't work, it's been reported that using the command "diskpart clean command" on Windows can fix this issue.
+
+
+
+### When booting a 3Node, how to fix the error: "no disks: registration failed"?
+
+There can be many different fixes for this error. Here are some troubleshooting tips to test separately:
+
+* In BIOS, enable AHCI
+* Make sure to [wipe the disks](../farmers/3node_building/4_wipe_all_disks.md) of the 3Nodes
+* If the 3Node has a RAID Controller:
+ * Disabled the RAID controller, OR;
+ * [Flash the RAID controller](https://fohdeesha.com/docs/perc.html) (i.e. crossflashing), OR;
+ * Change the controller to a Dell H310 controller (for Dell servers)
+* Try the command **badblocks** (replace **sda** with your specific disk). Note that this command will delete all the data on the disk
+ * ```
+ sudo badblocks -svw -b 512 -t 0x00 /dev/sda
+ ```
+
+
+### My SSD is sometimes detected as HDD by Zero-OS when there is a reboot. Is there a fix or a way to test the SSD disk?
+
+If your SSD disk shows as HDD, usually you can reboot the 3Node and Z-OS adjusts correctly.
+
+Anyone experiencing frequently this issue where Z-OS sometimes detects an SSD as HDD can try the following:
+
+* Boot up a live Ubuntu Desktop image
+* Run the benchmark utility within the Disks app
+* Check if the seektime of the disk is sufficient for Z-OS
+ * If the seektime is above 0.5ms, Z-OS will consider your SSD as HDD
+
+**Detailed Steps:**
+
+* Boot a Ubuntu Linux live USB
+* Install **gnome-disks** if it isn't already installed:
+ * ```
+ sudo apt install gnome-disks
+ ```
+* Open the application launcher and search for **Disks**
+* Select your disk
+* Click on the tree dots menu
+* Select **Benchmark Disk...**
+ * Use the default parameters
+ * **Transfer rate**: This is not relevant for our current test
+ * You can set to it to minimum (e.g. 2)
+ * **Sample size**: 10 MB is sufficient
+* Check the average access time on the [ThreeFold repository](https://www.github.com/threefoldtech/seektime)
+ * Check seek time for HDD and SSD
+ * A SSD needs to be <=0.5ms
+* If the result is above 0.5ms, this is why Z-OS doesn't recognize the disk properly
+ * You can then run diagnostics (e.g. smartmontools)
+ * If this is not fixable, you should change disk (e.g. take a more performing disk)
+
+Note: The green dots on the output represent seektime and that's what Z-OS is looking at. Specifically, it checks that the average seektime is below 0.5ms. If the seektime is above this, Z-OS will consider your SSD as HDD.
+
+
+
+### When booting a 3Node, I get the message: failed to register node: failed to create node: failed to submit extrinsic: Invalid Transaction: registration failed. What could fix this?
+
+The most probable fix to this error is simply to properly wipe your disk(s):
+
+* [Wipe your disks on Linux](#what-can-you-do-to-zero-out-your-disks-how-can-i-wipe-the-disks-of-my-3node-server-with-linux)
+
+* [Wipe your disks on Windows](#how-can-i-wipe-a-disk-with-windows)
+
+
+### I try to boot a 3Node, but I get the message no route with default gateway found. What does it mean?
+
+First, let's see the main terms. Default gateway act as an access point to other networks, in this case the TF Grid, when there is a back and forth exchange of data packets.
+
+While the last question implied a communication problem from the Grid, this error message usually means that the 3Node has communication problem. In short, it has difficulty reaching the TF Grid. There are many ways to troubleshoot this error. First, let's give the most direct solution. Make sure you have a direct connection with your Internet Service Provider (ISP): your 3Node should be connected to a router or a switch via an ethernet cable. Wifi doesn't work. Make sure your DHCP is set correctly.
+
+If the problem persists, check the default gateway of your 3Node and then make sure your router can reach it.
+
+*See next Q+A for more a possible solution.
+
+
+
+### I have trouble connecting the 3Node to the Grid with a 10GB NIC card. What can I do?
+
+As of now, Zero-OS sometimes has trouble with 10GB NIC card. The easiest solution to this is to connect your 3Node with the 1GB NIC card. This should solve the issue. More fine tuning might be needed to get your 3Node to work with a 10GB NIC card. Future Zero-OS version might solve this issue.
+
+
+
+### I switch the ethernet cable to a different port when my 3Node was running. Internet connection is lost. What can I do?
+
+When your 3Node boots, Zero-OS marks the NIC port. This means you cannot change NIC port when your 3Node is running. You can either put back the ethernet cable in the initial NIC port, or reboot the 3Node. At boot, Zero-OS will marks the new NIC port as the main entry.
+
+
+
+### I get the error Certificate is not yet valid when booting my 3Node server, what can I do?
+
+Make sure your firmware is up to date. If necessary, reinstall it. You might have to install then re-install the firmware if your system is very old.
+
+
+
+### When running wipefs to wipe my disks on Linux, I get either of the following errors: "syntax error near unexpected token" or "Probing Initialized Failed". Is there a fix?
+
+Many different reasons can cause this issue. When you get that error, sometimes it is because your are trying to wipe your boot USB by accident. If this is not the case, and you really are trying to wipe the correct disk, here are some fixes to try out, with the disk `sda` as an example:
+
+* Fix 1:
+ * Force the wiping of the disk:
+ * ```
+ sudo wipefs -af /dev/sda
+ ```
+* Fix 2:
+ * Unmount the disk then wipe it:
+ * ```
+ sudo umount /dev/sda
+ ```
+ * ```
+ sudo wipefs -a /dev/sda
+ ```
+
+
+
+### I did a format on my SSD disk, but Zero-OS still does not recognize them. What's wrong?
+
+Formatting is one thing, but to boot properly, Zero-OS needs to work on a completely wipe disk. Thus, make sure you [wipe your disks](#what-can-you-do-to-zero-out-your-disks-how-can-i-wipe-the-disks-with-linux). Formatting is not enough.
+
+
+
+### I have a Dell Rx10 server (R610, 710, 910). When I boot Zero-OS I get the message Probing EDD and the 3Node doesn't boot from there. What can I do?
+
+For the R610 and 710, you can simply re-flash the card. See [this link](https://fohdeesha.com/docs/perc.html) for more information. For the 910, you can’t re-flash the card. In this case, get a LSI Dell card and it should work. (They are cheap when you buy them used online.)
+
+
+### My 3Node doesn't boot properly without a monitor plugged in. What can I do?
+
+First, try to disable the "Halt On" mode in BIOS. If you do not have this option, try simply enabling the Legacy Support (Dell BIOS for example). If this doesn't work, try to plug in a Dummy Plug/Headless Ghost/Display Emulator in your 3Node. This will simulate a plugged monitor. This should fix the problem.
+
+
+### My 3Node is running on the Grid, but when I plugged in the monitor, it states: Disabling IR #16. Is there a problem?
+
+In general, you can simply ignore this error statement. This error is linked to the Nvidia binary driver. It simply means that your 3Node lost connection with the graphic card (by unplugging and replugging the monitor for example).
+
+
+### My 3Node won't boot without disabling the Secure Boot option, is it safe?
+
+In the case where you want to boot Zero-OS, disabling Secure Boot option is safe. With Secure Boot disabled, it can be easier or even necessary when it comes to booting Zero-OS. Secure Boot is used when you want to lock the BIOS/UEFI settings.
+
+
+
+### When I tried to boot my 3Node, at some point the screen went black, with or without a blinking hyphen or dash. What could cause this and what could I do to resolve the issue?
+
+There is a possibility that this happens because you are booting your 3Node on a HDD. A 3Node needs a minimum of 500GB of SSD to work properly.
+
+Also, make sure that you are using the correct boot option (Legacy BIOS or UEFI) in the Settings and that it corresponds to the correct booting image on the ThreeFold Bootstrap page.
+
+This problem often arises when you plugged your disks in the wrong controller. For example, try unpluging the disks from the SAS controller, and plug them in the SATA controller. Also, disable the SAS controller if needed.
+
+In a Legacy BIOS boot, make sure Legacy is enabled and disable *Data Execution Prevention* if possible.
+
+Also, it might have to do with your RAID controller configuration. Make sure this is properly set. For example, configuring all the HDD disks into one logical disk can fix this problem, or re-flashing the RAID card can also help.
+
+
+
+### My 3Nodes go offline after a modem reboot. Is there a way to prevent this?
+
+Yes, there are many ways to prevent this. An easy solution is to set the DHCP server to reserve local IPs for the 3Nodes MAC addresses.
+
+This problem is also preventable if your router stays online during the modem reboot.
+
+Indeed, rebooting the 3Nodes is necessary when there are local IP changes, as 3Nodes are addressed a local IP addresses when they are booted.
+
+The DHCP will addresses any local IP address that is available when you are booting a 3Node. Reserving local IP addresses is a good TF farming practice.
+
+
+
+### When I boot my 3Node, it reaches the Welcome to Zero-OS window, but it doesn't boot properly and there's an error message: failed to load object : type substrate..., what can I do?
+
+Usually simply rebooting the 3Node fixes this problem.
+
+
+
+### When I try to access iDRAC on a web browswer, even with protected mode off, I get the error The webpage cannot be found, what can I do?
+
+Open iDRAC in the Internet Explorer emulator extension (IE Tab) in Chrome, then update iDRAC. It should work elsewhere then. Sometimes, it will be needed to add "ST1=code" at the end of the IE Tab url.
+
+
+
+### When booting the 3Node, I get the error Network interface detected but autoconfiguration failed. What can I do?
+
+First make sure your network cable is plugged in and that your DHCP is working and responding. If you change the NIC port of the ethernet cable, make sure to reboot the 3Node so Zero-OS can change the NIC port attribution.
+
+Some farmers reported that this got fixed by simply powering off the 3Node(s), the router and modem for 2 minutes then powering it all back on. Resetting the modem and router (switch on the hardware) in the process can also help.
+
+If this doesn't work, try to upgrade the firmware of the NIC and the motherboard. If this still doesn't work, the NIC card might be broken. Try with another NIC card.
+
+
+
+### When I boot my Dell server, I get the message: All of the disks from your previous configuration are gone... Press any key to continue or 'C' to load the configuration utility. What can I do?
+
+Many changes to your server can lead to this message.
+
+Usually, the easiest solution is to reset the disk configuration in iDRAC's configuration utility.
+
+What can causes this message:
+
+1. During a new installation, the cables connecting to your external storage are not wired to the correct ports.
+2. Your RAID adapter has failed.
+3. Your SAS cables are not plugged properly or are malfunctioning.
+
+Note: Resetting the configuration will destroy all data on all virtual disks. Make sure you know what you are doing! In doubt, ask the TF community.
+
+
+
+### I have a Dell R620. In Zero-OS, I get the failure message No network card found and then the 3Node reebots after few seconds. The same happens for every LAN input. What can I do?
+
+The first thing to try here is to boot the server in BIOS mode instead of UEFI mode. If this doesn't fix the problem, try the following.
+
+Sometimes, this happens when the firmwares of BIOS, iDRAC, Lifecycle Controller and NIC are incompatible to each other. The solution is then to update them all correctly. Some problems can arise in the process.
+
+First, you should try to do the updates using iDRAC as you can update both iDRAC and BIOS there. If this does not work, try to update separate, with a live-linux distro, the BIOS, iDRAC and Lifecycle Controller. Once this is done, the server should be able to do a liveupdate, https to dell support website, via lifecycle-controller. This would update the other components. For more details on this method, watch [this video](https://www.youtube.com/watch?v=ISA7j2BKgjI).
+
+Note: Some farmers have reported that the Broadcom NIC card does not work well for Zero-OS and that a standard Intel PCI NIC card replacement resolved the issue. This could be a more straightforward method if updating the firmwares doesn't resolve the issue.
+
+
+
+### I am using freeDos to crossflash my raid controller on a Dell server, but I can't see the RAID controller with the Command Info. What can I do?
+
+Turn on the raid controller in the BIOS, otherwise freeDos does not show you the raid controller with the command Info.
+
+
+
+### Can I use a VGA to HDMI adaptor to connect a TV screen or monitor to the 3Node? I tried to boot a 3Node with a VGA to HDMI adaptor but the boot fails, what can I do?
+
+This might work, but it has been reported by farmers that Zero-OS might have difficulties booting when this is done with a VGA/HDMI adaptor on a TV screen. This is most likely due to the TV screen not supporting the output once the system loaded into Zero-OS. The easy fix to this issue is to use a standard computer monitor with a VGA plug.
+
+
+
+### When I try to boot my 3Node, the fans start spinning fast with a loud noise and the screen is black. What can I do to resolve this?
+
+There may be several causes to this issue. You can try to remove all the RAM sticks, to clean the dust and then to reseat the RAM sticks. If it still doesn't resolve the issue, you can check the RAM sticks one by one to see if one is malfunctioning. This often resolves the issue. Also, some cables might not be properly connected.
+
+
+
+### When booting Zero-OS with IPV6 configurations, I get the errors (1) dial tcp: address IPV6-address too many columns in address and (2) no pools matches key: not routable. What can I do to fix this issue?
+
+This usually means that the IPV6 attributed is not valid. It is also often caused when the DNS configuration does not resolve IPV6 correctly.
+
+To fix this issue, it is often necessary to adjust the IPV6 settings related to the router and the modem. Confirming with your Internet service provider (ISP) that the IPV6 settings are properly configured could also be necessary to fix the issue.
+
+
+
+
+### When booting a 3Node, Zero-OS downloads fine, but then I get the message: error no route with default gateway found, and the message: info check if interface has a cable plugged in. What could fix this?
+
+Make sure you have network stack enabled in BIOS. If so, check you ethernet port and make sure that it's clean. Also make sure the ethernet rj45 connectors are clean on both ends. If that does not work, verify the state of your SATA cables. If all this doesn't work, download and re-install Zero-OS.
+
+
+
+### How can I update Dell and HP servers to Intel E5-2600v2, E5-2400v2 and E5-4600v2, when applicable?
+
+There are many ressources online with steps on how to do this. You can check this [youtube video](https://www.youtube.com/watch?v=duzrULLtonM) on Dell and HP servers, as welll as this [documentation](https://ixnfo.com/en/hp-proliant-gen8-update-to-support-cpu-e5-2600v2-e5-2400v2-e5-4600v2.html) for HP Proliant Gen8.
+
+
+
+### How can I update the firmware and driver of a Dell PowerEdge server?
+
+Dell has excellen documentation for this. Read [this](https://www.dell.com/support/kbdoc/en-us/000128194/updating-firmware-and-drivers-on-dell-emc-poweredge-servers) for the detailed steps.
+
+
+
+### When I boot a 3Node in UEFI mode, it gets stuck at: Initializing Network Device, is there a way to fix this?
+
+In short, booting the 3Node in BIOS mode instead of UEFI mode usually fixes this issue.
+
+You can make bootable USB with the USB option of the [Zero-OS bootstrap image page](https://bootstrap.grid.tf/). Make sure to boot your server using BIOS and not UEFI. In the boot sequence, make the USB as your first choice to boot.
+
+
+
+### When I boot my 3Node, it gets stuck during the Zero-OS download. It never reaches 100%. What can I do to fix this issue?
+
+Here are some ways to troubleshoot your 3Node when it cannot download Zero-OS completely (to 100%):
+
+* Sometimes, just rebooting the 3Node and/or trying a little bit later can work.
+* It can help to reboot the modem and the router.
+* Make sure your BIOS/UEFI is up to date. Updating the BIOS/UEFI can help.
+* It can also help to set the correct date and time.
+
+
+
+### When booting a 3Node, I get the error=“context deadline exceeded” module=network error=failed to initialize rmb api failed to initialized admin mw: failed to get farm: farm not found: object not found. What can I do to fix this issue?
+
+Usually, the simple fix to this issue is to make sure that your bootstrap image is on the same network as your farm. For example, if you created your farm on the Main net, you should use a Main net Zero-OS bootstrap image.
+
+
+
+## ThreeFold Grid and Data
+
+
+### How is the farming minting reward calculated? Is the Grid always monitoring my 3Node?
+
+The Grid uses an algorithm that does not continually monitor the 3Node. It does its best to determine uptime through occasional checking in, which we call ping. The 3Node sends a ping to the Grid and the Grid sends a ping to the 3Node to confirm the reception. (Ping-Pong!)
+
+It’s helpful to understand that the Grid is really just 3Nodes and TF Chain (which itself is a collection of nodes). Nodes report their uptime by writing an entry on TF Chain, about once every two hours. These reports are used for minting.
+
+
+
+### How does communication happen on the ThreeFold Grid at the 3Node's level?
+
+There are two ways to get information about nodes. Once is to query TF Chain, and the other is to communicate with nodes directly.
+
+
+
+### What is the ThreeFold Node Status bot Telegram link?
+
+The link is the following: https://t.me/tfnodestatusbot.
+
+
+
+### How does the ThreeFold Node Status bot work? How can I use the ThreeFold Node Status bot to verify if my 3Node is online?
+
+1. Click on this link: https://t.me/tfnodestatusbot
+
+2. To subscribe a node, write the command */subscribe nodeID*. With the ID 100 as example, write: */subscribe 100*
+
+3. To verify the status of all your 3Nodes, write: */status*
+
+4. To verify the status of all your 3Nodes through Yggdrasil, write: */ping*
+
+Note: The bot should send you alerts when it considers any registered node to be offline.
+
+
+
+### How does the Telegram Status Bot get information from my 3Node? My 3Node is online on the ThreeFold Node Finder, but offline on the Telegram Status Bot, is this normal?
+
+The status bot communicates directly by sending pings to the nodes over Yggdrasil every five minutes. Therefore, it will report on temporary network interruptions that might not affect your total uptime calculation as used for minting.
+
+
+
+### I noticed that when I reboot my 3Node, the uptime counter on the ThreeFold Node Finder goes back to zero. Does it mean I lose uptime and the uptime start over again when I reboot the 3Node?
+
+No. The only uptime you lose is the time your 3Node was offline from the ThreeFold Grid. This ThreeFold Grid still has the data of your total uptime of the month. The Node Finder only shows this statistics as: "*This node has been up non-stop without being rebooted for now* [insert time]". If you maintain a total uptime above the minimum uptime, you're fine.
+
+*For now the farming rewards are proportional to the total uptime.
+
+
+### One of my nodes is showing the wrong location. Any problem with that?
+The ThreeFold Node Finder is showing your ISP location. This is perfectly normal.
+
+
+## Memory
+
+### Can I use different type of RAM for the same 3Node?
+
+No. Always use the same type of RAM per 3Node. If you use RDIMM, go all RDIMM, etc. Check your hardware specifications to make sure you have the right type of memory.
+
+
+
+### How can I know if the memory I am buying is correct for my specific hardware?
+
+To be sure, look into the owner's manual of your specific computer.
+
+In general, you can go to [Memory.net](https://memory.net/) and look for your specific computer model. As general steps, select your computer's system in *By system*, then select the series and then select the specific model of the series. You will then see available memories to buy from memory.net. You can also simply read the documentation at the bottom. The memory type supported by your computer will be explained. Then you can buy the memory needed from any other computer store.
+
+For servers, you can check with Cloudninja's documentation [here](https://cloudninjas.com/pages/server-memory). Search for your specific hardware and look for the compatible memory. This reference is good for rack and tower servers.
+
+
+
+### What do the terms RDIMM, LDIMM, UDIMM, LRDIMM, FBDIMM mean when it comes to RAM memory sticks?
+
+Well first, the DIMM means dual inline memory module.
+
+* U stands for or unregistered (or unbuffered).
+
+* R stands for registered memory.
+
+* LR stands for load-reduced.
+
+* FB stands for fully-buffered.
+
+
+
+### What is the difference between ECC and non-ECC memory?
+
+ECC means error correction code memory. This type of memory can detect and correct data corruption. Non-ECC mostly cannot detect nor correct, but some can detect, but never correct data corruption. Check your hardware specifications to make sure you have the right type of memory (ECC or non-ECC).
+
+
+### How can I change the RAM memory sticks on my 3Nodes? How can I achieve dual channel configuration with sticks of RAM?
+
+First, always use RAM sticks of the same size and type. It should be noted on your motherboard which slots to populate first. As a general guide, there is usually 2 slots A and B, with each 2 memory stick entries. You must then install the ram sticks on A1 and B1 in order to achieve dual channel, then A2 and B2 if you have more (visual order: A1 A2 B1 B2).
+
+> Example: You want to start with your largest sticks, evenly distributed between both processors and work your way down to your smallest. Let's take an example with 2 processors as well as 4x 16GB sticks and 4x 8GB sticks. The arrangement would be A1-16GB, B1-16GB, A2-16GB, B2-16GB, A3-8GB, B3-8GB, A4-8GB, B4-8GB. Avoid odd numbers as well. You optimally want pairs. So if you only have 5x 8GB sticks, only install 4 until you have an even 6.
+
+
+
+### What does RAM mean?
+
+RAM means random access memory. Those type of memory can be read and changed in any order.
+
+
+
+
+### What does DIMM mean when it comes to RAM sticks?
+
+It means *dual in-line memory module*. This type of computer memory is natively 64 bits, enabling fast data transfer.
+
+
+
+### I have 24 DIMMS ram slots on my server. Can I use them all?
+
+Be careful when installing memory on a server. Always check your server's documentation to make sure your RAM sticks combination are correct.
+
+For example, on the Dell R720, you can have 24x16gb RAM ECC sticks, but it can only handle 16 Quad ranked DIMMs. In this case, you can fill up all slots with registered DIMMs if you have a maximum of 4 quad DIMMS ranked on each CPU.
+
+# Ask a Question to the ThreeFold Community
+
+If you have any question, you can ask the ThreeFold community via the ThreeFold forum or the ThreeFold Telegram channels:
+
+* [ThreeFold Forum](https://forum.threefold.io/)
+* [ThreeFold General TG Channel](https://t.me/threefold)
+* [ThreeFold Farmer TG Channel](https://t.me/threefoldfarmers)
+* [TF Grid Tester TG Channel](https://t.me/threefoldtesting)
+
+> Note 1: If we wrote something wrong, tell us!
+
+> Note 2: This is a collective effort. A big *Thank You* to the great ThreeFold Community. Many Q+A are contributions from the enthusiast farmers, users and developers of the ever-growing ThreeFold community.
\ No newline at end of file
diff --git a/collections/documentation/faq/img/3nodes.png b/collections/documentation/faq/img/3nodes.png
new file mode 100644
index 0000000..c7fe437
Binary files /dev/null and b/collections/documentation/faq/img/3nodes.png differ
diff --git a/collections/documentation/faq/img/faq_img_readme.md b/collections/documentation/faq/img/faq_img_readme.md
new file mode 100644
index 0000000..5c71297
--- /dev/null
+++ b/collections/documentation/faq/img/faq_img_readme.md
@@ -0,0 +1 @@
+# Image folder for the FAQ of the Threefold Manual 3.0
diff --git a/collections/documentation/faq/img/minting2022.png b/collections/documentation/faq/img/minting2022.png
new file mode 100644
index 0000000..3335f69
Binary files /dev/null and b/collections/documentation/faq/img/minting2022.png differ
diff --git a/collections/documentation/faq/img/tf_general.jpg b/collections/documentation/faq/img/tf_general.jpg
new file mode 100644
index 0000000..9925a1d
Binary files /dev/null and b/collections/documentation/faq/img/tf_general.jpg differ
diff --git a/collections/documentation/faq/img/tf_grid.png b/collections/documentation/faq/img/tf_grid.png
new file mode 100644
index 0000000..3996369
Binary files /dev/null and b/collections/documentation/faq/img/tf_grid.png differ
diff --git a/collections/documentation/faq/img/tf_grid_3nodes.png b/collections/documentation/faq/img/tf_grid_3nodes.png
new file mode 100644
index 0000000..49f5a98
Binary files /dev/null and b/collections/documentation/faq/img/tf_grid_3nodes.png differ
diff --git a/collections/documentation/faq/img/wethreepedia_developer.png b/collections/documentation/faq/img/wethreepedia_developer.png
new file mode 100644
index 0000000..58a1967
Binary files /dev/null and b/collections/documentation/faq/img/wethreepedia_developer.png differ
diff --git a/collections/documentation/faq/img/wethreepedia_faq_poster.jpg b/collections/documentation/faq/img/wethreepedia_faq_poster.jpg
new file mode 100644
index 0000000..bf43918
Binary files /dev/null and b/collections/documentation/faq/img/wethreepedia_faq_poster.jpg differ
diff --git a/collections/documentation/faq/img/wethreepedia_validator.jpg b/collections/documentation/faq/img/wethreepedia_validator.jpg
new file mode 100644
index 0000000..f23a7cf
Binary files /dev/null and b/collections/documentation/faq/img/wethreepedia_validator.jpg differ
diff --git a/collections/documentation/farmers/3node_building/1_create_farm.md b/collections/documentation/farmers/3node_building/1_create_farm.md
new file mode 100644
index 0000000..460fc32
--- /dev/null
+++ b/collections/documentation/farmers/3node_building/1_create_farm.md
@@ -0,0 +1,88 @@
+ 1. Create a Farm
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Create a TFChain Account](#create-a-tfchain-account)
+- [Create a Farm](#create-a-farm)
+- [Create a ThreeFold Connect Wallet](#create-a-threefold-connect-wallet)
+- [Add a Stellar Address for Payout](#add-a-stellar-address-for-payout)
+ - [Farming Rewards Distribution](#farming-rewards-distribution)
+- [More Information](#more-information)
+
+***
+
+## Introduction
+
+We cover the basic steps to create a farm with the ThreeFold Dashboard. We also create a TFConnect app wallet to receive the farming rewards.
+
+## Create a TFChain Account
+
+We create a TFChain account using the ThreeFold Dashboard.
+
+Go to the [ThreeFold Dashboard](https://dashboard.grid.tf/), click on **Create Account**, choose a password and click **Connect**.
+
+![tfchain_create_account](./img/dashboard_tfchain_create_account.png)
+
+Once your profile gets activated, you should find your Twin ID and Address generated under your Mnemonics for verification. Also, your Account Balance will be available at the top right corner under your profile name.
+
+![tf_mnemonics](./img/dashboard_tf_mnemonics.png)
+
+## Create a Farm
+
+We create a farm using the dashboard.
+
+In the left-side menu, select **Farms** -> **Your Farms**.
+
+![your_farms](./img/dashboard_your_farms.png)
+
+Click on **Create Farm**, choose a farm name and then click **Create**.
+
+![create_farm](./img/dashboard_create_farm.png)
+
+![farm_name](./img/dashboard_farm_name.png)
+
+## Create a ThreeFold Connect Wallet
+
+Your farming rewards should be sent to a Stellar wallet with a TFT trustline enabled. The simplest way to proceed is to create a TF Connect app wallet as the TFT trustline is enabled by default on this wallet. For more information on TF Connect, read [this section](../../threefold_token/storing_tft/tf_connect_app.md).
+
+Let's create a TFConnect Wallet and take note of the wallet address. First, download the app.
+
+This app is available for [Android](https://play.google.com/store/apps/details?id=org.jimber.threebotlogin&hl=en&gl=US) and [iOS](https://apps.apple.com/us/app/threefold-connect/id1459845885).
+
+- Note that for Android phones, you need at minimum Android Nougat, the 8.0 software version.
+- Note that for iOS phones, you need at minimum iOS 14.5. It will be soon available to iOS 13.
+
+Open the app, click **SIGN UP**, choose a ThreeFold Connect Id, write your email address, take note of the seed phrase and choose a pin. Once this is done, you will have to verify your email address. Check your email inbox.
+
+In the app menu, click on **Wallet** and then click on **Create Initial Wallet**.
+
+To find your wallet address, click on the **circled i** icon at the bottom of the screen.
+
+![dashboard_tfconnect_wallet_1](./img/dashboard_tfconnect_wallet_1.png)
+
+Click on the button next to your Stellar address to copy the address.
+
+![dashboard_tfconnect_wallet_2](./img/dashboard_tfconnect_wallet_2.png)
+
+You will need the TF Connect wallet address for the next section.
+
+> Note: Make sure to keep your TF Connect Id and seed phrase in a secure place offline. You will need these two components to recover your account if you lose access.
+
+## Add a Stellar Address for Payout
+
+In the **Your Farms** section of the dashboard, click on **Add/Edit Stellar Payout Address**.
+
+![dashboard_walletaddress_1](./img/dashboard_walletaddress_1.png)
+
+Paste your Stellar wallet address and click **Submit**.
+
+![dashboard_walletaddress_2](./img/dashboard_walletaddress_2.png)
+
+### Farming Rewards Distribution
+
+Farming rewards will be sent to your farming wallet around the 8th of each month. This can vary depending on the situation. The minting is done automatically by code and verified by humans as a double check.
+
+## More Information
+
+For more information, such as setting IP addresses, you can consult the [Dashboard Farms section](../../dashboard/farms/farms.md).
\ No newline at end of file
diff --git a/collections/documentation/farmers/3node_building/2_bootstrap_image.md b/collections/documentation/farmers/3node_building/2_bootstrap_image.md
new file mode 100644
index 0000000..9234242
--- /dev/null
+++ b/collections/documentation/farmers/3node_building/2_bootstrap_image.md
@@ -0,0 +1,177 @@
+ 2. Create a Zero-OS Bootstrap Image
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Download the Zero-OS Bootstrap Image](#download-the-zero-os-bootstrap-image)
+- [Burn the Zero-OS Bootstrap Image](#burn-the-zero-os-bootstrap-image)
+ - [CD/DVD BIOS](#cddvd-bios)
+ - [USB Key BIOS+UEFI](#usb-key-biosuefi)
+ - [BalenaEtcher (MAC, Linux, Windows)](#balenaetcher-mac-linux-windows)
+ - [CLI (Linux)](#cli-linux)
+ - [Rufus (Windows)](#rufus-windows)
+- [Additional Information (Optional)](#additional-information-optional)
+ - [Expert Mode](#expert-mode)
+ - [Use a Specific Kernel](#use-a-specific-kernel)
+ - [Disable GPU](#disable-gpu)
+ - [Bootstrap Image URL](#bootstrap-image-url)
+ - [Zeros-OS Bootstrapping](#zeros-os-bootstrapping)
+ - [Zeros-OS Expert Bootstrap](#zeros-os-expert-bootstrap)
+
+***
+
+## Introduction
+
+We will now learn how to create a Zero-OS bootstrap image in order to boot a DIY 3Node.
+
+## Download the Zero-OS Bootstrap Image
+
+Let's download the Zero-OS bootstrap image.
+
+In the Farms section of the Dashboard, click on **Bootstrap Node Image**
+
+![dashboard_bootstrap_farm](./img/dashboard_bootstrap_farm.png)
+
+or use the direct link [https://v3.bootstrap.grid.tf](https://v3.bootstrap.grid.tf):
+
+```
+https://v3.bootstrap.grid.tf
+```
+
+![Farming_Create_Farm_21](./img/farming_createfarm_21.png)
+
+This is the Zero-OS v3 Bootstrapping page.
+
+![Farming_Create_Farm_22](./img/farming_createfarm_22.png)
+
+Write your farm ID and choose production mode.
+
+![Farming_Create_Farm_23](./img/farming_createfarm_23.png)
+
+If your system is new, you might be able to run the bootstrap in UEFI mode.
+
+![Farming_Create_Farm_24](./img/farming_createfarm_24.png)
+
+For older systems, run the bootstrap in BIOS mode. For BIOS CD/DVD, choose **ISO**. For BIOS USB, choose **USB**
+
+Download the bootstrap image. Next, we will burn the bootstrap image.
+
+
+
+## Burn the Zero-OS Bootstrap Image
+
+We show how to burn the Zero-OS bootstrap image. A quick and modern way is to burn the bootstrap image on a USB key.
+
+### CD/DVD BIOS
+
+For the BIOS **ISO** image, download the file and burn it on a DVD.
+
+### USB Key BIOS+UEFI
+
+There are many ways to burn the bootstrap image on a USB key. The easiest way that works for all operating systems is to use BalenaEtcher. We also provide other methods.
+
+#### BalenaEtcher (MAC, Linux, Windows)
+
+For **MAC**, **Linux** and **Windows**, you can use [BalenaEtcher](https://www.balena.io/etcher/) to load/flash the image on a USB stick. This program also formats the USB in the process. This will work for the option **EFI IMG** for UEFI boot, and with the option **USB** for BIOS boot. Simply follow the steps presented to you and make sure you select the bootstrap image file you downloaded previously.
+
+> Note: There are alternatives to BalenaEtcher (e.g. [usbimager](https://gitlab.com/bztsrc/usbimager/)).
+
+**General Steps with BalenaEtcher:**
+
+1. Download BalenaEtcher
+2. Open BalenaEtcher
+3. Select **Flash from file**
+4. Find and select the bootstrap image (with your correct farm ID)
+5. Select **Target** (your USB key)
+6. Select **Flash**
+
+That's it. Now you have a bootstrap image on Zero-OS as a bootable removable media device.
+
+
+#### CLI (Linux)
+
+For the BIOS **USB** and the UEFI **EFI IMG** images, you can do the following on Linux:
+
+ sudo dd status=progress if=FILELOCATION.ISO(or .IMG) of=/dev/sd*
+
+Here the * is to indicate that you must adjust according to your disk. To see your disks, write lsblk in the command window. Make sure you select the proper disk!
+
+*If you USB key is not new, make sure that you format it before burning the Zero-OS image.
+
+#### Rufus (Windows)
+
+For Windows, if you are using the "dd" able image, instead of writing command line, you can use the free USB flashing program called [Rufus](https://sourceforge.net/projects/rufus.mirror/) and it will automatically do this without needing to use the command line. Rufus also formats the boot media in the process.
+
+## Additional Information (Optional)
+
+We cover some additional information. Note that the following information is not needed for a basic farm setup.
+
+### Expert Mode
+
+You can use the [expert mode](https://v3.bootstrap.grid.tf/expert) to generate specific Zero-OS bootstrap images.
+
+Along the basic options of the normal bootstrap mode, the expert mode allows farmers to add extra kernel arguments and decide which kernel to use from a vast list of Zero-OS kernels.
+
+#### Use a Specific Kernel
+
+You can use the expert mode to choose a specific kernel. Simply set the information you normally use and then select the proper kernel you need in the **Kernel** drop-down list.
+
+![](./img/bootstrap_kernel_list.png)
+
+#### Disable GPU
+
+You can use the expert mode to disable GPU on your 3Node.
+
+![](./img/bootstrap_disable-gpu.png)
+
+In the expert mode of the Zero-OS Bootstrap generator, fill in the following information:
+
+- Farmer ID
+ - Your current farm ID
+- Network
+ - The network of your farm
+- Extra kernel arguments
+ - ```
+ disable-gpu
+ ```
+- Kernel
+ - Leave the default kernel
+- Format
+ - Choose a bootstrap image format
+- Click on **Generate**
+- Click on **Download**
+
+### Bootstrap Image URL
+
+In both normal and expert mode, you can use the generated URL to quickly download a Zero-OS bootstrap image based on your farm specific setup.
+
+Using URLs can be a very quick and efficient way to create new bootstrap images once your familiar with the Zero-OS bootstrap URL template and some potential varations.
+
+```
+https://.bootstrap.grid.tf//////.../
+```
+
+Note that the arguments and the kernel are optional.
+
+The following content will provide some examples.
+
+#### Zeros-OS Bootstrapping
+
+On the [main page](https://v3.bootstrap.grid.tf/), once you've written your farm ID and selected a network, you can copy the generated URL of any given image format.
+
+For example, the following URL is a download link to an **EFI IMG** of the Zero-OS bootstrap image of farm 1 on the main TFGrid v3 network:
+
+```
+https://v3.bootstrap.grid.tf/uefimg/prod/1
+```
+
+#### Zeros-OS Expert Bootstrap
+
+You can use the generated sublink at the **Generate step** of the expert mode to get a quick URL to download your bootstrap image.
+
+- After setting the parameters and arguments, click on **Generate**
+- Add the **Target** content to the following URL `https://v3.bootstrap.grid.tf`
+ - For example, the following URL sets an **ipxe** script of the Zero-OS bootstrap of farm 1 on the main TFGrid v3 network, with the **disable-gpu** function enabled as an extra kernel argument and a specific kernel:
+ - ```
+ https://v3.bootstrap.grid.tf/ipxe/test/1/disable-gpu/zero-os-development-zos-v3-generic-b8706d390d.efi
+ ```
\ No newline at end of file
diff --git a/collections/documentation/farmers/3node_building/3_set_hardware.md b/collections/documentation/farmers/3node_building/3_set_hardware.md
new file mode 100644
index 0000000..6f053dd
--- /dev/null
+++ b/collections/documentation/farmers/3node_building/3_set_hardware.md
@@ -0,0 +1,188 @@
+ 3. Set the Hardware
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Hardware Requirements](#hardware-requirements)
+ - [3Node Requirements Summary](#3node-requirements-summary)
+- [Bandwidth Requirements](#bandwidth-requirements)
+- [Link to Share Farming Setup](#link-to-share-farming-setup)
+- [Powering the 3Node](#powering-the-3node)
+ - [Surge Protector](#surge-protector)
+ - [Power Distribution Unit (PDU)](#power-distribution-unit-pdu)
+ - [Uninterrupted Power Supply (UPS)](#uninterrupted-power-supply-ups)
+ - [Generator](#generator)
+- [Connecting the 3Node to the Internet](#connecting-the-3node-to-the-internet)
+ - [Z-OS and Switches](#z-os-and-switches)
+- [Using Onboard Storage (3Node Servers)](#using-onboard-storage-3node-servers)
+- [Upgrading a DIY 3Node](#upgrading-a-diy-3node)
+
+***
+
+
+## Introduction
+
+In this section of the ThreeFold Farmers book, we cover the essential farming requirements when it comes to ThreeFold 3Node hardware.
+
+The essential information are available in the section [3Node Requirements Summary](#3node-requirements-summary).
+
+## Hardware Requirements
+
+
+You need a theoretical minimum of 500 GB of SSD and 2 GB of RAM on a mini pc, desktop or server. In short, for peak optimization, aim for 100 GB of SSD and 8GB of RAM per thread. (Thread is equivalent to virtual core or logical core.)
+
+Also, TFDAO might implement a farming parameter based on [passmark](https://www.cpubenchmark.net/cpu_list.php). From the ongoing discussion on the Forum, you should aim at a CPU mark of 1000 and above per core.
+
+> 3Node optimal farming hardware ratio -> 100 GB of SSD + 8 GB of RAM per Virtual Core
+
+Note that you can run Zero-OS on a Virtual Machine (VM), but you won't farm any TFT from this process. To farm TFT, Zero-OS needs to be on bare metal.
+
+Also, note that ThreeFold runs its own OS, which is Zero-OS. You thus need to start with completely wiped disks. You cannot farm TFT with Windows, Linux or MAC OS installed on your disks. If you need to use such OS temporarily, boot it in Try mode with a removable media (USB key).
+
+Note: Once you have the necessary hardware, you need to [create a farm](./1_create_farm.md), [create a Zero-OS bootstrap image](./2_bootstrap_image.md), [wipe your disks](./4_wipe_all_disks.md) and [set the BIOS/UEFI](./5_set_bios_uefi.md) . Then you can [boot your 3Node](./6_boot_3node.md). If you are planning in building a farm in data center, [read this section](../advanced_networking/advanced_networking_toc.md).
+
+
+
+### 3Node Requirements Summary
+
+
+
+Any computer with the following specifications can be used as a DIY 3Node.
+
+- Any 64-bit hardware with an Intel or AMD processor chip.
+- Servers, desktops and mini computers type hardware are compatible.
+- A minimum of 500 GB of SSD and a bare minimum of 2 GB of RAM is required.
+- A ratio of 100GB of SSD and 8GB of RAM per thread is recommended.
+- A wired ethernet connection is highly recommended to maximize reliability and the ability to farm TFT.
+- A [passmark](https://www.passmark.com/) of 1000 per core is recommended and will probably be a minimum requirement in the future.
+
+*A passmark of 1000 per core is recommend and will be a minimum requirement in the future. This is not yet an official requirement. A 3Node with less than 1000 passmark per core of CPU would not be penalized if it is registered before the DAO settles the [Passmark Question](https://forum.threefold.io/t/cpu-benchmarking-for-reward-calculations/2479).
+
+
+
+## Bandwidth Requirements
+
+
+
+A 3Node connects to the ThreeFold Grid and transfers information, whether it is in the form of compute, storage or network units (CU, SU, NU respectively). The more resources your 3Nodes offer to the Grid, the more bandwidth will be needed to transfer the additional information. In this section, we cover general guidelines to make sure you have enough bandwidth on the ThreeFold Grid when utilization will be happening.
+
+Note that the TFDAO will need to discuss and settle on clearer guidelines in the near future. For now, we propose those general guidelines. Being aware of these numbers as you build and scale your ThreeFold farm will set you in the proper direction.
+
+> **The strict minimum for one Titan is 1 mbps of bandwidth**.
+
+If you want to expand your ThreeFold farm, you should check the following to make sure your bandwidth will be sufficient when there will be Grid utilization.
+
+**Bandwidth per 3Node Equation**
+
+> min Bandwidth per 3Node (mbps) = 10 * max((Total SSD TB / 1 TB),(Total Threads / 8 Threads),(Total GB / 64 GB)) + 10 * (Total HDD TB / 2)
+
+This equation means that for each TB of HDD you need 5 mbps of bandwidth, and for each TB of SSD, 8 Threads and 64GB of RAM (whichever is higher), you need 10 mbps of bandwidth.
+
+This means a proper bandwidth for a Titan would be 10 mbps. As stated, 1 mbps is the strict minimum for one Titan.
+
+
+
+## Link to Share Farming Setup
+
+
+If you want ideas and suggestions when it comes to building DIY 3Nodes, a good place to start is by checking what other farmers have built. [This post on the Forum](https://forum.threefold.io/t/lets-share-our-farming-setup/286) is a great start. The following section also contains great DIY 3Node ideas.
+
+## Powering the 3Node
+
+### Surge Protector
+
+A surge protector is highly recommended for your farm and your 3Nodes. This ensures your 3Nodes will not overcharge if a power surge happens. Whole-house surge protectors are also an option.
+
+### Power Distribution Unit (PDU)
+
+A PDU (power distribution unit) is useful in big server settings in order to manage your wattage and keep track of your power consumption.
+
+
+### Uninterrupted Power Supply (UPS)
+
+
+A UPS (uninterrupted power supply) is great for a 3Node if your power goes on and off frequently for short periods of time. This ensures your 3Node does not need to constantly reboot. If your electricity provider is very reliable, a UPS might not be needed, as the small downtime resulting from rare power outages with not exceed the DIY downtime limit*. (95% uptime, 5% downtime = 36 hours per month.) Of course, for greater Grid utilization experience, considering adding a UPS to your ThreeFold farm can be highly beneficial.
+
+Note: Make sure to have AC Recovery Power set properly so your 3Node goes back online if power shutdowns momentarily. UPS are generally used in data center to make sure people have enough time to do a "graceful" shutdown of the units when power goes off. In the case of 3Nodes, they do not need graceful shutdowns as Zero-OS cannot lose data while functioning. The only way to power down a 3Node is simply to turn it off directly on the machine.
+
+
+### Generator
+
+
+A generator will be needed for very large installation with or without an unsteady main power supply.
+
+
+
+## Connecting the 3Node to the Internet
+
+As a general consideration, to connect a 3Node to the Internet, you must use an Ethernet cable and set DHCP as a network management protocol. Note that WiFi is not supported with ThreeFold farming.
+
+The general route from the 3Node to the Internet is the following:
+
+> 3Node -> Switch (optional) -> Router -> Modem
+
+Note that most home routers come with a built-in switch to provide multiple Ethernet ports. Using a stand-alone switch is optional, but can come quite handy when farmers have many 3Nodes.
+
+
+
+### Z-OS and Switches
+
+Switches can be managed or unmanaged. Managed switches come with managed features made available to the user (typically more of such features on premium models).
+
+Z-OS can work with both types of switches. As long as there's a router reachable on the other end offering DHCP and a route to the public internet, it's not important what's in between. Generally speaking, switches are more like cables, just part of the pipes that connect devices in a network.
+
+We present a general overview of the two types of switches.
+
+**Unmanaged Switches**
+
+Unmanaged are the most common type and if someone just says "switch" this is probably what they mean. These switches just forward traffic along to its destination in a plug and play manner with no configuration. When a switch is connected to a router, you can think of the additional free ports on the switch as essentially just extra ports on the router. It's a way to expand the available ports and sometimes also avoid running multiple long cables. My nodes are far from my router, so I run a single long ethernet cable to a switch next to the nodes and then use multiple shorter cables to connect from the switch to the nodes.
+
+**Managed Switches**
+
+Managed switches have more capabilities than unmanaged switches and they are not very common in home settings (at least not as standalone units). Some of our farmers do use managed switches. These switches offer much more control and also require configuration. They can enable advanced features like virtual LANs to segment the network.
+
+
+
+## Using Onboard Storage (3Node Servers)
+
+If your 3Node is based on a server, you can either use PCIe slots and PCIe-NVME adapter to install SSD NVME disk, or you can use the onboard storage.
+
+Usually, servers use RAID technology for onboard storage. RAID is a technology that has brought resilience and security to the IT industry. But it has some limitations that ThreeFold did not want to get stuck with. ThreeFold developed a different and more efficient way to [store data reliably](https://library.threefold.me/info/threefold#/cloud/threefold__cloud_products?id=storage-quantum-safe-filesystem). This Quantum Safe Storage overcomes some of the shortfalls of RAID and is able to work over multiple nodes geographically spread on the TF Grid. This means that there is no RAID controller in between data storage and the TF Grid.
+
+For your 3Nodes, you want to bypass RAID in order for Zero-OS to have bare metals on the system.
+
+To use onboard storage on a server without RAID, you can
+
+1. [Re-flash](https://fohdeesha.com/docs/perc.html) the RAID card
+2. Turn on HBA/non-RAID mode
+3. Install a different card.
+
+For HP servers, you simply turn on the HBA mode (Host Bus Adapter).
+
+For Dell servers, you can either cross, or [re-flash](https://fohdeesha.com/docs/perc.html), the RAID controller with an “IT-mode-Firmware” (see this [video](https://www.youtube.com/watch?v=h5nb09VksYw)) or get a DELL H310-controller (which has the non-RAID option). Otherwise, you can install a NVME SSD with a PCIe adaptor, and turn off the RAID controller.
+
+
+
+Once the disks are wiped, you can shutdown your 3Node and remove the Linux Bootstrap Image (USB key). Usually, there will be a message telling you when to do so.
+
+
+
+## Upgrading a DIY 3Node
+
+
+
+As we've seen in the [List of Common DIY 3Nodes](#list-of-common-diy-3nodes), it is sometimes necessary, and often useful, to upgrade your hardware.
+
+**Type of upgrades possible**
+
+- Add TBs of SSD/HDD
+- Add RAM
+- Change CPU
+- Change BIOS battery
+- Change fans
+
+For some DIY 3Node, no upgrades are required and this constitutes a good start if you want to explore DIY building without going into too much additional steps.
+
+For in-depth videos on how to upgrade mini-pc and rack servers, watch these great [DIY videos](https://www.youtube.com/user/floridanelson).
\ No newline at end of file
diff --git a/collections/documentation/farmers/3node_building/3node_building.md b/collections/documentation/farmers/3node_building/3node_building.md
new file mode 100644
index 0000000..8f26b69
--- /dev/null
+++ b/collections/documentation/farmers/3node_building/3node_building.md
@@ -0,0 +1,14 @@
+ Building a DIY 3Node
+
+This section of the ThreeFold Farmers book presents the necessary and basic steps to build a DIY 3Node.
+
+For advanced farming information, such as GPU farming and room parameters, refer to the section [Farming Optimization](../farming_optimization/farming_optimization.md).
+
+ Table of Contents
+
+- [1. Create a Farm](./1_create_farm.md)
+- [2. Create a Zero-OS Bootstrap Image](./2_bootstrap_image.md)
+- [3. Set the Hardware](./3_set_hardware.md)
+- [4. Wipe All the Disks](./4_wipe_all_disks.md)
+- [5. Set the BIOS/UEFI](./5_set_bios_uefi.md)
+- [6. Boot the 3Node](./6_boot_3node.md)
\ No newline at end of file
diff --git a/collections/documentation/farmers/3node_building/4_wipe_all_disks.md b/collections/documentation/farmers/3node_building/4_wipe_all_disks.md
new file mode 100644
index 0000000..4e252a5
--- /dev/null
+++ b/collections/documentation/farmers/3node_building/4_wipe_all_disks.md
@@ -0,0 +1,106 @@
+ 4. Wipe All the Disks
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Main Steps](#main-steps)
+- [1. Create a Linux Bootstrap Image](#1-create-a-linux-bootstrap-image)
+- [2. Boot Linux in *Try Mode*](#2-boot-linux-in-try-mode)
+- [3. Use wipefs to Wipe All the Disks](#3-use-wipefs-to-wipe-all-the-disks)
+- [Troubleshooting](#troubleshooting)
+
+***
+
+## Introduction
+
+In this section of the ThreeFold Farmers book, we explain how to wipe all the disks of your 3Node.
+
+
+
+## Main Steps
+
+It only takes a few steps to wipe all the disks of a 3Node.
+
+1. Create a Linux Bootstrap Image
+2. Boot Linux in *Try Mode*
+3. Wipe All the Disks
+
+ThreeFold runs its own OS, which is Zero-OS. You thus need to start with completely wiped disks. Note that ALL disks must be wiped. Otherwise, Zero-OS won't boot.
+
+An easy method is to simply download a Linux distribution and wipe the disk with the proper command line in the Terminal.
+
+We will show how to do this with Ubuntu 20.04. LTS. This distribution is easy to use and it is thus a good introduction for Linux, in case you haven't yet explored this great operating system.
+
+
+
+## 1. Create a Linux Bootstrap Image
+
+Download the Ubuntu 20.04 ISO file [here](https://releases.ubuntu.com/20.04/) and burn the ISO image on a USB key. Make sure you have enough space on your USB key. You can also use other Linux Distro such as [GRML](https://grml.org/download/), if you want a lighter ISO image.
+
+The process here is the same as in section [Burning the Bootstrap Image](./2_bootstrap_image.md#burn-the-zero-os-bootstrap-image), but with the Linux ISO instead of the Zero-OS ISO. [BalenaEtcher](https://www.balena.io/etcher/) is recommended as it formats your USB in the process, and it is available for MAC, Windows and Linux.
+
+
+
+## 2. Boot Linux in *Try Mode*
+
+When you boot the Linux ISO image, make sure to choose *Try Mode*. Otherwise, it will install Linux on your computer. You do not want this.
+
+
+
+## 3. Use wipefs to Wipe All the Disks
+
+When you use wipefs, you are removing all the data on your disk. Make sure you have no important data on your disks, or make sure you have copies of your disks before doing this operation, if needed.
+
+Once Linux is booted, go into the terminal and write the following command lines.
+
+First, you can check the available disks by writing in a terminal or in a shell:
+
+```
+lsblk
+```
+
+To see what disks are connected, write this command:
+
+```
+fdisk -l
+```
+
+If you want to wipe one specific disk, here we use *sda* as an example, write this command:
+
+```
+sudo wipefs -a /dev/sda
+```
+
+And replace the "a" in sda by the letter of your disk, as shown when you did *lsblk*. The term *sudo* gives you the correct permission to do this.
+
+To wipe all the disks in your 3Node, write the command:
+
+```
+sudo for i in /dev/sd*; do wipefs -a $i; done
+```
+
+If you have any `fdisk` entries that look like `/dev/nvme`, you'll need to adjust the command line.
+
+For a nvme disk, here we use *nvme0* as an example, write:
+
+```
+sudo wipefs -a /dev/nvme0
+```
+
+And replace the "0" in nvme0 by the number corresponding to your disk, as shown when you did *lsblk*.
+
+To wipe all the nvme disks, write this command line:
+
+```
+sudo for i in /dev/nvme*; do wipefs -a $i; done
+```
+
+## Troubleshooting
+
+If you're having issues wiping the disks, you might need to use **--force** or **-f** with wipefs (e.g. **sudo wipefs -af /dev/sda**).
+
+If you're having trouble getting your disks recognized by Zero-OS, some farmers have had success enabling AHCI mode for SATA in their BIOS.
+
+If you are using a server with onboard storage, you might need to [re-flash the RAID card](../../faq/faq.md#is-there-a-way-to-bypass-raid-in-order-for-zero-os-to-have-bare-metals-on-the-system-no-raid-controller-in-between-storage-and-the-grid).
+
+
diff --git a/collections/documentation/farmers/3node_building/5_set_bios_uefi.md b/collections/documentation/farmers/3node_building/5_set_bios_uefi.md
new file mode 100644
index 0000000..4c6c3e9
--- /dev/null
+++ b/collections/documentation/farmers/3node_building/5_set_bios_uefi.md
@@ -0,0 +1,172 @@
+ 5. Set the BIOS/UEFI
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Z-OS and DHCP](#z-os-and-dhcp)
+ - [Regular Computer and 3Node Network Differences](#regular-computer-and-3node-network-differences)
+ - [Static IP Addresses](#static-ip-addresses)
+- [The Essential Features of BIOS/UEFI for a 3Node](#the-essential-features-of-biosuefi-for-a-3node)
+- [Setting the Remote Management of a Server with a Static IP Address (Optional)](#setting-the-remote-management-of-a-server-with-a-static-ip-address-optional)
+- [Update the BIOS/UEFI firmware (Optional)](#update-the-biosuefi-firmware-optional)
+ - [Check the BIOS/UEFI version on Windows](#check-the-biosuefi-version-on-windows)
+ - [Check the BIOS/UEFI version on Linux](#check-the-biosuefi-version-on-linux)
+ - [Update the BIOS firmware](#update-the-bios-firmware)
+- [Additional Information](#additional-information)
+ - [BIOS/UEFI and Zero-OS Bootstrap Image Combinations](#biosuefi-and-zero-os-bootstrap-image-combinations)
+ - [Troubleshoot](#troubleshoot)
+
+
+***
+
+## Introduction
+
+In this section of the ThreeFold Farmers book, we explain how to properly set the BIOS/UEFI of your 3Node.
+
+Note that the BIOS mode is usually needed for older hardware while the UEFI mode is usually needed for newer hardware, when it comes to booting properly Zero-OS on your DIY 3Node.
+
+If it doubt, start with UEFI and if it doesn't work as expected, try with BIOS.
+
+Before diving into the BIOS/UEFI settings, we will present some general considerations on Z-OS and DHCP.
+
+## Z-OS and DHCP
+
+The operating system running on the 3Nodes is called Zero-OS (Z-OS). When it comes to setting the proper network for your 3Node farm, you must use DHCP since Z-OS is going to request an IP from the DHCP server if there's one present, and it won't get network connectivity if there's no DHCP.
+
+The Z-OS philosophy is to minimize configuration wherever possible, so there's nowhere to supply a static config when setting your 3Node network. Instead, the farmer is expected to provide DHCP.
+
+While it is possible to set fixed IP addresses with the DHCP for the 3Nodes, it is recommended to avoid this and just set the DHCP normally without fixed IP addresses.
+
+By setting DHCP in BIOS/UEFI, an IP address is automatically assigned by your router to your 3Node every time you boot it.
+
+### Regular Computer and 3Node Network Differences
+
+For a regular computer (not a 3Node), if you want to use a static IP in a network with DHCP, you'd first turn off DHCP and then set the static IP to an IP address outside the DHCP range. That being said, with Z-OS, there's no option to turn off DHCP and there's nowhere to set a static IP, besides public config and remote management. In brief, the farmer must provide DHCP, either on a private or a public range, for the 3Node to boot.
+
+### Static IP Addresses
+
+In the ThreeFold ecosystem, there are only two situations where you would work with static IP addresses: to set a public config to a 3Node or a farm, and to remotely manage your 3Nodes.
+
+**Static IP and Public Config**
+
+You can [set a static IP for the public config of a 3Node or a farm](./1_create_farm.md#optional-add-public-ip-addresses). In thise case, the 3Node takes information from TF Chain and uses it to set a static configuration on a NIC (or on a virtual NIC in the case of single NIC systems).
+
+**Static IP and Remote Management**
+
+You can [set a static IP address to remotely manage a 3Node](#setting-the-remote-management-of-a-server-static-ip-address).
+
+
+
+## The Essential Features of BIOS/UEFI for a 3Node
+
+There are certain things that you should make sure are set properly on your 3Node.
+
+As a general advice, you can Load Defaults (Settings) on your BIOS, then make sure the options below are set properly.
+
+* Choose the correct combination of BIOS/UEFI and bootstrap image on [https://bootstrap.grid.tf/](https://bootstrap.grid.tf/)
+ * Newer system will use UEFI
+ * Older system will use BIOS
+ * Hint: If your 3Node boot stops at *Initializing Network Devices*, try the other method (BIOS or UEFI)
+* Set Multi-Processor and Hyperthreading at Enabled
+ * Sometimes, it will be written Virtual Cores, or Logical Cores.
+* Set Virtualization at Enabled
+ * On Intel, it is denoted as CPU virtualization and on ASUS, it is denoted as SVM.
+ * Make sure virtualization is enabled and look for the precise terms in your specific BIOS/UEFI.
+* Set AC Recovery at Last Power State
+ * This will make sure your 3Node restarts after losing power momentarily.
+* Select the proper Boot Sequence for the 3Node to boot Zero-OS from your bootstrap image
+ * e.g., if you have a USB key as a bootstrap image, select it in Boot Sequence
+* Set Server Lookup Method (or the equivalent) at DNS. Only use Static IP if you know what you are doing.
+ * Your router will assign a dynamic IP address to your 3Node when it connects to Internet.
+* Set Client Address Method (or the equivalent) at DHCP. Only use Static IP if you know what you are doing.
+ * Your router will assign a dynamic IP address to your 3Node when it connects to Internet.
+* Secure Boot should be left at disabled
+ * Enable it if you know what you are doing. Otherwise, it can be set at disabled.
+
+
+
+
+## Setting the Remote Management of a Server with a Static IP Address (Optional)
+
+
+Note from the list above that by enabling the DHCP and DNS in BIOS, dynamic IP addresses will be assigned to 3Nodes. This way, you do not need any specific port configuration when booting a 3Node.
+
+As long as the 3Node is connected to the Internet via an ethernet cable (WiFi is not supported), Zero-OS will be able to boot. By setting DHCP in BIOS, an IP address is automatically assigned to your 3Node every time you boot it. This section concerns 3Node servers with remote management functions and interfaces.
+
+You can set up a node through static routing at the router without DHCP by assigning the MAC address of the NIC to a IP address within your private subnet. This will give a static IP address to your 3Node.
+
+With a static IP address, you can then configure remote management on servers. For Dell, [iDRAC](https://www.dell.com/support/kbdoc/en-us/000134243/how-to-setup-and-manage-your-idrac-or-cmc-for-dell-poweredge-servers-and-blades) is used, and for HP, [ILO](https://support.hpe.com/hpesc/public/docDisplay?docId=a00045463en_us&docLocale=en_US) is used.
+
+
+
+## Update the BIOS/UEFI firmware (Optional)
+
+
+Updating the BIOS firmware is not always necessary, but to do so can help prevent future errors and troubleshootings. Making sure the Date and Time are set correctly can also help the booting process.
+
+Note: updating the BIOS/UEFI firmware is optional, but recommended.
+
+
+### Check the BIOS/UEFI version on Windows
+
+Hit *Start*, type in *cmd* in the search box and click on *Command Prompt*. Write the line
+
+> wmic bios get smbiosbiosversion
+
+This will give you the BIOS or UEFI firmware of your PC.
+
+### Check the BIOS/UEFI version on Linux
+
+Simply type the following command
+
+> sudo dmidecode | less
+
+or this line:
+
+> sudo dmidecode -s bios-version
+
+### Update the BIOS firmware
+
+1. On the manufacturer's website, download the latest BIOS/UEFI firmware
+2. Put the file on a USB flash drive (+unzip if necessary)
+3. Restart your hardware and enter the BIOS/UEFI settings
+4. Navigate the menus to update the BIOS/UEFI
+
+## Additional Information
+
+### BIOS/UEFI and Zero-OS Bootstrap Image Combinations
+
+To properly boot the Zero-OS image, you can either use an image made for a BIOS system or a UEFI system, this depends on your system.
+
+BIOS is older technology. It means *Basic Input/Output System*.
+
+UEFI is newer technology. It means *Unified Extensible Firmware Interface*. BIOS/UEFI is, in a way, the link between the hardware and the software of your computer.
+
+In general, setting a 3Node is similar whether it is with a BIOS or UEFI system. The important is to choose the correct combination of boot media and boot mode (BIOS/UEFI).
+
+The bootstrap images are available [here](https://bootstrap.grid.tf/).
+
+The choices are:
+
+1. EFI IMG - UEFI
+2. EFI FILE - UEFI
+3. iPXE - Boot from network
+4. ISO - BIOS
+5. USB - BIOS
+6. LKRN - Boot from network
+
+Choices 1 and 2 are for UEFI (newer models).
+Choices 4 and 5 are for BIOS (newer models).
+Choices 3 and 6 are mainly for network boot.
+
+Refer to [this previous section](./2_bootstrap_image.md) for more information on creating a Zero-OS bootstrap image.
+
+For information on how to boot Zero-OS with iPXE, read [this section](./6_boot_3node.md#advanced-booting-methods-optional).
+
+### Troubleshoot
+
+You might have to try UEFI first and if it doesn't work, try BIOS. Usually when this is the case (UEFI doesn't work with your current computer), the following message will be shown:
+
+> Initializing Network Devices...
+
+And then... nothing. This means that you are still in the BIOS of the hardware and boot is not even started yet. When this happens, try the BIOS mode of your computer.
\ No newline at end of file
diff --git a/collections/documentation/farmers/3node_building/6_boot_3node.md b/collections/documentation/farmers/3node_building/6_boot_3node.md
new file mode 100644
index 0000000..404e2e9
--- /dev/null
+++ b/collections/documentation/farmers/3node_building/6_boot_3node.md
@@ -0,0 +1,169 @@
+ 6. Boot the 3Node
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [1. Booting the 3Node with Zero-OS](#1-booting-the-3node-with-zero-os)
+- [2. Check the 3Node Status Online](#2-check-the-3node-status-online)
+- [3. Receive the Farming Rewards](#3-receive-the-farming-rewards)
+- [Advanced Booting Methods (Optional)](#advanced-booting-methods-optional)
+ - [PXE Booting with OPNsense](#pxe-booting-with-opnsense)
+ - [PXE Booting with pfSense](#pxe-booting-with-pfsense)
+- [Booting Issues](#booting-issues)
+ - [Multiple nodes can run with the same node ID](#multiple-nodes-can-run-with-the-same-node-id)
+
+***
+
+
+## Introduction
+
+We explain how to boot the 3Node with the Zero-OS bootstrap image with a USB key. We also include optional advanced booting methods using OPNSense and pfSense.
+
+One of the great features of Zero-OS is that it can be completely run within the cache of your 3Node. Indeed, the booting device that contains your farm ID will connect to the ThreeFold Grid and download everything needed to run smoothly. There are many benefits in terms of security and protection of data that comes with this.
+
+## 1. Booting the 3Node with Zero-OS
+
+To boot Zero-OS, insert your Zero-OS bootstrap image USB key, power on your computer and choose the right booting sequence and parameters ([BIOS or UEFI](./5_set_bios_uefi.md)) in your BIOS/UEFI settings. Then, restart the 3Node. Zero-OS should boot automatically.
+
+Note that you need an ethernet cable connected to your router or switch. You cannot farm on the ThreeFold Grid with Wifi.
+
+The first time you boot a 3Node, it will be written: “This node is not registered (farmer : NameOfFarm). This is normal. The Grid will create a node ID and you will be able to see it on screen. This can take a couple of minutes.
+
+If time passes (an hour and more) and the node does not get registered, in many cases, [wiping the disks](./4_wipe_all_disks.md) all over again and trying another reboot usually resolves this issue.
+
+Once you have your node ID, you can also go on the ThreeFold Dashboard to see your 3Node and verify that your 3Node is online.
+
+## 2. Check the 3Node Status Online
+
+You can use the ThreeFold [Node Finder](../../dashboard/deploy/node_finder.md) to verify that your 3Node is online.
+
+* [ThreeFold Main Net Dashboard](https://dashboard.grid.tf/)
+* [ThreeFold Test Net Dashboard](https://dashboard.test.grid.tf/)
+* [ThreeFold Dev Net Dashboard](https://dashboard.dev.grid.tf/)
+* [ThreeFold QA Net Dashboard](https://dashboard.qa.grid.tf/)
+
+
+## 3. Receive the Farming Rewards
+
+The farming reward will be sent once per month at the address you gave when you set up your farm. You can review this process [here](./1_create_farm.md#add-a-stellar-address-for-payout).
+
+That's it. You've now completed the necessary steps to build a DIY 3Node and to connect it to the Grid.
+
+## Advanced Booting Methods (Optional)
+
+### PXE Booting with OPNsense
+
+> This documentation comes from the [amazing Network Booting Guide](https://forum.ThreeFold.io/t/network-booting-tutorial/2688) by @Fnelson on the ThreeFold Forum.
+
+Network booting ditches your standard boot USB with a local server. This TFTP server delivers your boot files to your 3 nodes. This can be useful in bigger home farms, but is all but mandatory in a data center setup.
+
+Network boot setup is quite easy and is centered about configuring a TFTP server. There are essentially 2 options for this, a small dedicated server such as a raspberry pi, or piggybacking on your pfsense or opnsense router. I would recommend the latter as it eliminates another piece of equipment and is probably more reliable.
+
+**Setting Up Your Router to Allow Network Booting**
+
+These steps are for OPNsense, PFsense may differ. These set are required regardless of where you have your TFTP server.
+
+> Services>DHCPv4>LAN>Network Booting
+
+Check “Enable Network Booting”
+
+Enter the IP address of your TFTP server under “Set next-server IP”. This may be the router’s IP or whatever device you are booting from.
+
+Enter “pxelinux.0” under Set default bios filename.
+
+Ignore the TFTP Server settings.
+
+
+**TFTP server setup on a debian machine such as Ubuntu or Raspberry Pi**
+
+> apt-get update
+>
+> apt-get install tftpd-hpa
+>
+> cd /srv/tftp/
+>
+> wget http://ftp.nl.debian.org/debian/dists/buster/main/installer-amd64/current/images/netboot/netboot.tar.gz
+>
+> wget http://ftp.nl.debian.org/debian/dists/buster/main/installer-amd64/current/images/netboot/pxelinux.0
+>
+> wget https://bootstrap.grid.tf/krn/prod/ --no-check-certificate
+>
+> mv ipxe-prod.lkrn
+>
+> tar -xvzf netboot.tar.gz
+>
+> rm version.info netboot.tar.gz
+>
+> rm pxelinux.cfg/default
+>
+> chmod 777 /srv/tftp/pxelinux.cfg (optional if next step fails)
+>
+> echo 'default ipxe-prod.lkrn' >> pxelinux.cfg/default
+
+
+**TFTP Server on a OPNsense router**
+
+> Note: When using PFsense instead of OPNsense, steps are probably similar, but the directory or other small things may differ.
+
+The first step is to download the TFTP server plugin. Go to system>firmware>Status and check for updates, follow prompts to install. Then click the Plugins tab and search for tftp, install os-tftp. Once that is installed go to Services>TFTP (you may need to refresh page). Check the Enable box and input your router ip (normally 192.168.1.1). Click save.
+
+Turn on ssh for your router. In OPNsense it is System>Settings>Administration. Then check the Enable, root login, and password login. Hop over to Putty and connect to your router, normally 192.168.1.1. Login as root and input your password. Hit 8 to enter the shell.
+
+In OPNsense the tftp directory is /usr/local/tftp
+
+> cd /usr/local
+>
+> mkdir tftp
+>
+> cd ./tftp
+>
+> fetch http://ftp.nl.debian.org/debian/dists/buster/main/installer-amd64/current/images/netboot/netboot.tar.gz
+>
+> fetch http://ftp.nl.debian.org/debian/dists/buster/main/installer-amd64/current/images/netboot/pxelinux.0
+>
+> fetch https://bootstrap.grid.tf/krn/prod/
+>
+> mv ipxe-prod.lkrn
+>
+> tar -xvzf netboot.tar.gz
+>
+> rm version.info netboot.tar.gz
+>
+> rm pxelinux.cfg/default
+>
+> echo 'default ipxe-prod.lkrn' >> pxelinux.cfg/default
+
+You can get out of shell by entering exit or just closing the window.
+
+**3Node Setup**
+
+Set the server to BIOS boot and put PXE or network boot as the first choice. At least on Dell machines, make sure you have the network cable in plug 1 or it won’t boot.
+
+
+
+### PXE Booting with pfSense
+
+> This documentation comes from the [amazing Network Booting Guide](https://forum.threefold.io/t/network-booting-tutorial/2688/7) by @TheCaptain on the ThreeFold Forum.
+
+These are the steps required to enable PXE booting on pfSense. This guide assumes you’ll be using the router as your PXE server; pfSense allows boot file uploads directly from its web GUI.
+
+* Log into your pfSense instance
+ * Go to System>Package Manager
+ * Search and add ‘tftpd’ package under ‘Available Packages’ tab
+* Go to Services>TFTP Server
+ * Under ‘Settings’ tab check enable and enter the router IP in TFTP Server Bind IP field
+* Switch to ‘Files’ tab under Services>TFTP Server and upload your ‘ipxe-prod.efi’ file acquired from https://v3.bootstrap.grid.tf/ (second option labeled ‘EFI Kernel’)
+* Go to Services>DHCP Server
+ * Under ‘Other Options’ section click Display Advance next to ‘TFTP’ and enter router IP
+ * Click Display Advance next to ‘Network Booting’
+ * Check enable, enter router IP in Next Server field
+ * Enter ipxe-prod.efi in Default BIOS file name field
+
+That's it! You’ll want to ensure your clients are configured with boot priority set as IPv4 in first spot. You might need to disable secure boot and enable legacy boot within BIOS.
+
+## Booting Issues
+
+### Multiple nodes can run with the same node ID
+
+This is a [known issue](https://github.com/threefoldtech/info_grid/issues/122) and will be resolved once the TPM effort gets finalized.
+
diff --git a/collections/documentation/farmers/3node_building/gpu_farming.md b/collections/documentation/farmers/3node_building/gpu_farming.md
new file mode 100644
index 0000000..b684163
--- /dev/null
+++ b/collections/documentation/farmers/3node_building/gpu_farming.md
@@ -0,0 +1,72 @@
+GPU Farming
+
+Welcome to the *GPU Farming* section of the ThreeFold Manual!
+
+In this guide, we delve into the realm of GPU farming, shedding light on the significance of Graphics Processing Units (GPUs) and how they can be seamlessly integrated into the ThreeFold ecosystem.
+
+Table of Contents
+
+- [Understanding GPUs](#understanding-gpus)
+- [Get Started](#get-started)
+- [Install the GPU](#install-the-gpu)
+- [GPU Node and the Farmerbot](#gpu-node-and-the-farmerbot)
+- [Set a Price for the GPU Node](#set-a-price-for-the-gpu-node)
+- [Check the GPU Node on the Node Finder](#check-the-gpu-node-on-the-node-finder)
+- [Reserving the GPU Node](#reserving-the-gpu-node)
+- [Questions and Feedback](#questions-and-feedback)
+
+***
+
+## Understanding GPUs
+
+A Graphics Processing Unit, or GPU, is a specialized electronic circuit designed to accelerate the rendering of images and videos. Originally developed for graphics-intensive tasks in gaming and multimedia applications, GPUs have evolved into powerful parallel processors with the ability to handle complex computations, such as 3D rendering, AI and machine learning.
+
+In the context of ThreeFold, GPU farming involves harnessing the computational power of Graphics Processing Units to contribute to the decentralized grid. This empowers users to participate in the network's mission of creating a more equitable and efficient internet infrastructure.
+
+## Get Started
+
+In this guide, we focus on the integration of GPUs with a 3Node, the fundamental building block of the ThreeFold Grid. The process involves adding a GPU to enhance the capabilities of your node, providing increased processing power and versatility for a wide range of tasks. Note that any Nvidia or AMD graphics card should work as long as it's supported by the system.
+
+## Install the GPU
+
+We cover the basic steps to install the GPU on your 3Node.
+
+* Find a proper GPU model for your specific 3Node hardware
+* Install the GPU on the server
+ * Note: You might need to move or remove some pieces of your server to make room for the GPU
+* (Optional) Boot the 3Node with a Linux distro (e.g. Ubuntu) and use the terminal to check if the GPU is recognized by the system
+ * ```
+ sudo lshw -C Display
+ ```
+ * Output example with an AMD Radeon (on the line `product: ...`)
+![gpu_farming](./img/cli_display_gpu.png)
+* Boot the 3Node with the ZOS bootstrap image
+
+## GPU Node and the Farmerbot
+
+If you are using the Farmerbot, it might be a good idea to first boot the GPU node without the Farmerbot (i.e. to remove the node in the config file and restart the Farmerbot). Once you've confirmed that the GPU is properly detected by TFChain, you can then put back the GPU node in the config file and restart the Farmerbot. While this is not necessary, it can be an effective way to test the GPU node separately.
+
+## Set a Price for the GPU Node
+
+You can [set additional fees](../farming_optimization/set_additional_fees.md) for your GPU dedicated node on the [TF Dashboard](https://dashboard.grid.tf/).
+
+When a user reserves your 3Node as a dedicated node, you will receive TFT payments once every 24 hours. These TFT payments will be sent to the TFChain account of your farm's twin.
+
+## Check the GPU Node on the Node Finder
+
+You can use the [Node Finder](../../dashboard/deploy/node_finder.md) on the [TF Dashboard](https://dashboard.grid.tf/) to verify that the node is displayed as having a GPU.
+
+* On the Dashboard, go to the Node Finder
+* Under **Node ID**, write the node ID of the GPU node
+* Once the results are displayed, you should see **1** under **GPU**
+ * If you are using the Status bot, you might need to change the node status under **Select Nodes Status** (e.g. **Down**, **Standby**) to see the node's information
+
+> Note: It can take some time for the GPU parameter to be displayed.
+
+## Reserving the GPU Node
+
+Now, users can reserve the node in the **Dedicated Nodes** section of the Dashboard and then deploy workloads using the GPU. For more information, read [this documentation](../../dashboard/deploy/dedicated_machines.md).
+
+## Questions and Feedback
+
+If you have any questions or feedback, we invite you to discuss with the ThreeFold community on the [ThreeFold Forum](https://forum.threefold.io/) or on the [ThreeFold Farmer chat](https://t.me/threefoldfarmers) on Telegram.
\ No newline at end of file
diff --git a/collections/documentation/farmers/3node_building/img/bootstrap_disable-gpu.png b/collections/documentation/farmers/3node_building/img/bootstrap_disable-gpu.png
new file mode 100644
index 0000000..f72f450
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/bootstrap_disable-gpu.png differ
diff --git a/collections/documentation/farmers/3node_building/img/bootstrap_kernel_list.png b/collections/documentation/farmers/3node_building/img/bootstrap_kernel_list.png
new file mode 100644
index 0000000..3f3f559
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/bootstrap_kernel_list.png differ
diff --git a/collections/documentation/farmers/3node_building/img/cli_display_gpu.png b/collections/documentation/farmers/3node_building/img/cli_display_gpu.png
new file mode 100644
index 0000000..777b68b
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/cli_display_gpu.png differ
diff --git a/collections/documentation/farmers/3node_building/img/dashboard_1.png b/collections/documentation/farmers/3node_building/img/dashboard_1.png
new file mode 100644
index 0000000..bcfed02
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/dashboard_1.png differ
diff --git a/collections/documentation/farmers/3node_building/img/dashboard_2.png b/collections/documentation/farmers/3node_building/img/dashboard_2.png
new file mode 100644
index 0000000..1f6538a
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/dashboard_2.png differ
diff --git a/collections/documentation/farmers/3node_building/img/dashboard_4.png b/collections/documentation/farmers/3node_building/img/dashboard_4.png
new file mode 100644
index 0000000..54c413a
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/dashboard_4.png differ
diff --git a/collections/documentation/farmers/3node_building/img/dashboard_5.png b/collections/documentation/farmers/3node_building/img/dashboard_5.png
new file mode 100644
index 0000000..8bfe4e4
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/dashboard_5.png differ
diff --git a/collections/documentation/farmers/3node_building/img/dashboard_6.png b/collections/documentation/farmers/3node_building/img/dashboard_6.png
new file mode 100644
index 0000000..980d6d3
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/dashboard_6.png differ
diff --git a/collections/documentation/farmers/3node_building/img/dashboard_bootstrap_farm.png b/collections/documentation/farmers/3node_building/img/dashboard_bootstrap_farm.png
new file mode 100644
index 0000000..82f8759
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/dashboard_bootstrap_farm.png differ
diff --git a/collections/documentation/farmers/3node_building/img/dashboard_create_farm.png b/collections/documentation/farmers/3node_building/img/dashboard_create_farm.png
new file mode 100644
index 0000000..858da74
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/dashboard_create_farm.png differ
diff --git a/collections/documentation/farmers/3node_building/img/dashboard_farm_name.png b/collections/documentation/farmers/3node_building/img/dashboard_farm_name.png
new file mode 100644
index 0000000..c250693
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/dashboard_farm_name.png differ
diff --git a/collections/documentation/farmers/3node_building/img/dashboard_tf_mnemonics.png b/collections/documentation/farmers/3node_building/img/dashboard_tf_mnemonics.png
new file mode 100644
index 0000000..ec92cde
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/dashboard_tf_mnemonics.png differ
diff --git a/collections/documentation/farmers/3node_building/img/dashboard_tfchain_create_account.png b/collections/documentation/farmers/3node_building/img/dashboard_tfchain_create_account.png
new file mode 100644
index 0000000..1bdcd95
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/dashboard_tfchain_create_account.png differ
diff --git a/collections/documentation/farmers/3node_building/img/dashboard_tfconnect_wallet_1.png b/collections/documentation/farmers/3node_building/img/dashboard_tfconnect_wallet_1.png
new file mode 100644
index 0000000..52a9679
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/dashboard_tfconnect_wallet_1.png differ
diff --git a/collections/documentation/farmers/3node_building/img/dashboard_tfconnect_wallet_2.png b/collections/documentation/farmers/3node_building/img/dashboard_tfconnect_wallet_2.png
new file mode 100644
index 0000000..1ef9dc9
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/dashboard_tfconnect_wallet_2.png differ
diff --git a/collections/documentation/farmers/3node_building/img/dashboard_walletaddress_1.png b/collections/documentation/farmers/3node_building/img/dashboard_walletaddress_1.png
new file mode 100644
index 0000000..da65f76
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/dashboard_walletaddress_1.png differ
diff --git a/collections/documentation/farmers/3node_building/img/dashboard_walletaddress_2.png b/collections/documentation/farmers/3node_building/img/dashboard_walletaddress_2.png
new file mode 100644
index 0000000..a36e095
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/dashboard_walletaddress_2.png differ
diff --git a/collections/documentation/farmers/3node_building/img/dashboard_your_farms.png b/collections/documentation/farmers/3node_building/img/dashboard_your_farms.png
new file mode 100644
index 0000000..58de47a
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/dashboard_your_farms.png differ
diff --git a/collections/documentation/farmers/3node_building/img/farming_001.png b/collections/documentation/farmers/3node_building/img/farming_001.png
new file mode 100644
index 0000000..b666fce
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/farming_001.png differ
diff --git a/collections/documentation/farmers/3node_building/img/farming_30.png b/collections/documentation/farmers/3node_building/img/farming_30.png
new file mode 100644
index 0000000..d810d13
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/farming_30.png differ
diff --git a/collections/documentation/farmers/3node_building/img/farming_createfarm_1.png b/collections/documentation/farmers/3node_building/img/farming_createfarm_1.png
new file mode 100644
index 0000000..95545d8
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/farming_createfarm_1.png differ
diff --git a/collections/documentation/farmers/3node_building/img/farming_createfarm_2.png b/collections/documentation/farmers/3node_building/img/farming_createfarm_2.png
new file mode 100644
index 0000000..488f970
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/farming_createfarm_2.png differ
diff --git a/collections/documentation/farmers/3node_building/img/farming_createfarm_21.png b/collections/documentation/farmers/3node_building/img/farming_createfarm_21.png
new file mode 100644
index 0000000..ab36d45
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/farming_createfarm_21.png differ
diff --git a/collections/documentation/farmers/3node_building/img/farming_createfarm_22.png b/collections/documentation/farmers/3node_building/img/farming_createfarm_22.png
new file mode 100644
index 0000000..7cfc6ba
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/farming_createfarm_22.png differ
diff --git a/collections/documentation/farmers/3node_building/img/farming_createfarm_23.png b/collections/documentation/farmers/3node_building/img/farming_createfarm_23.png
new file mode 100644
index 0000000..3537771
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/farming_createfarm_23.png differ
diff --git a/collections/documentation/farmers/3node_building/img/farming_createfarm_24.png b/collections/documentation/farmers/3node_building/img/farming_createfarm_24.png
new file mode 100644
index 0000000..4b2da96
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/farming_createfarm_24.png differ
diff --git a/collections/documentation/farmers/3node_building/img/farming_createfarm_3.png b/collections/documentation/farmers/3node_building/img/farming_createfarm_3.png
new file mode 100644
index 0000000..349ec17
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/farming_createfarm_3.png differ
diff --git a/collections/documentation/farmers/3node_building/img/farming_createfarm_4.png b/collections/documentation/farmers/3node_building/img/farming_createfarm_4.png
new file mode 100644
index 0000000..b717bfe
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/farming_createfarm_4.png differ
diff --git a/collections/documentation/farmers/3node_building/img/farming_createfarm_5.png b/collections/documentation/farmers/3node_building/img/farming_createfarm_5.png
new file mode 100644
index 0000000..a022a1f
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/farming_createfarm_5.png differ
diff --git a/collections/documentation/farmers/3node_building/img/farming_createfarm_6.png b/collections/documentation/farmers/3node_building/img/farming_createfarm_6.png
new file mode 100644
index 0000000..922b77b
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/farming_createfarm_6.png differ
diff --git a/collections/documentation/farmers/3node_building/img/farming_createfarm_7.png b/collections/documentation/farmers/3node_building/img/farming_createfarm_7.png
new file mode 100644
index 0000000..e4bdb6e
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/farming_createfarm_7.png differ
diff --git a/collections/documentation/farmers/3node_building/img/farming_createfarm_8.png b/collections/documentation/farmers/3node_building/img/farming_createfarm_8.png
new file mode 100644
index 0000000..e8d023c
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/farming_createfarm_8.png differ
diff --git a/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_1.png b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_1.png
new file mode 100644
index 0000000..c93f67c
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_1.png differ
diff --git a/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_10.png b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_10.png
new file mode 100644
index 0000000..880b4fe
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_10.png differ
diff --git a/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_11.png b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_11.png
new file mode 100644
index 0000000..f5aea83
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_11.png differ
diff --git a/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_12.png b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_12.png
new file mode 100644
index 0000000..afb06e0
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_12.png differ
diff --git a/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_13.png b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_13.png
new file mode 100644
index 0000000..110a97a
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_13.png differ
diff --git a/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_14.png b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_14.png
new file mode 100644
index 0000000..8c2d2c6
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_14.png differ
diff --git a/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_15.png b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_15.png
new file mode 100644
index 0000000..30d41f2
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_15.png differ
diff --git a/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_16.png b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_16.png
new file mode 100644
index 0000000..320536f
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_16.png differ
diff --git a/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_2.png b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_2.png
new file mode 100644
index 0000000..b1276d9
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_2.png differ
diff --git a/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_4.png b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_4.png
new file mode 100644
index 0000000..9ce84ae
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_4.png differ
diff --git a/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_5.png b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_5.png
new file mode 100644
index 0000000..488bad0
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_5.png differ
diff --git a/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_6.png b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_6.png
new file mode 100644
index 0000000..77b3f03
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_6.png differ
diff --git a/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_7.png b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_7.png
new file mode 100644
index 0000000..21e55fc
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_7.png differ
diff --git a/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_8.png b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_8.png
new file mode 100644
index 0000000..a00dfa1
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_8.png differ
diff --git a/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_9.png b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_9.png
new file mode 100644
index 0000000..64b9685
Binary files /dev/null and b/collections/documentation/farmers/3node_building/img/tf_dashboard_2023_9.png differ
diff --git a/collections/documentation/farmers/3node_building/minting_receipts.md b/collections/documentation/farmers/3node_building/minting_receipts.md
new file mode 100644
index 0000000..2b84f1b
--- /dev/null
+++ b/collections/documentation/farmers/3node_building/minting_receipts.md
@@ -0,0 +1,98 @@
+Minting Receipts
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Access the Reports](#access-the-reports)
+- [Available Information](#available-information)
+
+***
+
+## Introduction
+
+Once you have the receipt hash of your node minting, you can get the [minting report](../../dashboard/tfchain/tf_minting_reports.md) of your node.
+
+## Access the Reports
+
+- On the Dashboard, go to **TFChain** -> **TF Minting Reports**
+- Enter your receipt hash
+- Consult your minting report
+
+## Available Information
+
+The ThreeFold Alpha minting tool will present the following information for each minting receipt hash:
+
+- Node Info
+ - Node ID
+ - Farm Name and ID
+ - Measured Uptime
+- Node Resources
+ - CU
+ - SU
+ - NU
+ - CRU
+ - MRU
+ - SRU
+ - HRU
+- TFT Farmed
+- Payout Address
+
+
\ No newline at end of file
diff --git a/collections/documentation/farmers/advanced_networking/advanced_networking_toc.md b/collections/documentation/farmers/advanced_networking/advanced_networking_toc.md
new file mode 100644
index 0000000..f30a86f
--- /dev/null
+++ b/collections/documentation/farmers/advanced_networking/advanced_networking_toc.md
@@ -0,0 +1,13 @@
+ Advanced Networking
+
+Welcome to the *Advanced Networking* section of the ThreeFold Manual.
+
+In this section, we provide advanced networking tips for farms with public IPs and in data centers (DC). We also cover the differences between IPv4 and IPv6 networking.
+
+Table of Contents
+
+- [Networking Overview](./networking_overview.md)
+- [Network Considerations](./network_considerations.md)
+- [Network Setup](./network_setup.md)
+
+> Note: This documentation does not constitute a complete set of knowledge on setting farms with public IP addresses in a data center. Please make sure to do your own research and communicate with your data center and your Internet service provider for any additional information.
\ No newline at end of file
diff --git a/collections/documentation/farmers/advanced_networking/network_considerations.md b/collections/documentation/farmers/advanced_networking/network_considerations.md
new file mode 100644
index 0000000..0576fd9
--- /dev/null
+++ b/collections/documentation/farmers/advanced_networking/network_considerations.md
@@ -0,0 +1,120 @@
+Network Considerations
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Running ZOS (v2) at Home](#running-zos-v2-at-home)
+- [Running ZOS (v2) in a Multi-Node Farm in a DC](#running-zos-v2-in-a-multi-node-farm-in-a-dc)
+ - [Necessities](#necessities)
+ - [IPv6](#ipv6)
+ - [Routing/Firewalling](#routingfirewalling)
+ - [Multi-NIC Nodes](#multi-nic-nodes)
+ - [Farmers and the TFGrid](#farmers-and-the-tfgrid)
+
+***
+
+## Introduction
+
+Running ZOS on a node is just a matter of booting it with a USB stick, or with a dhcp/bootp/tftp server with the right configuration so that the node can start the OS.
+Once it starts booting, the OS detects the NICs, and starts the network configuration. A Node can only continue it's boot process till the end when it effectively has received an IP address and a route to the Internet. Without that, the Node will retry indefinitely to obtain Internet access and not finish it's startup.
+
+So a Node needs to be connected to a __wired__ network, providing a dhcp server and a default gateway to the Internet, be it NATed or plainly on the public network, where any route to the Internet, be it IPv4 or IPv6 or both is sufficient.
+
+For a node to have that ability to host user networks, we **strongly** advise to have a working IPv6 setup, as that is the primary IP stack we're using for the User Network's Mesh to function.
+
+## Running ZOS (v2) at Home
+
+Running a ZOS Node at home is plain simple. Connect it to your router, plug it in the network, insert the preconfigured USB stick containing the bootloader and the `farmer_id`, power it on.
+
+## Running ZOS (v2) in a Multi-Node Farm in a DC
+
+Multi-Node Farms, where a farmer wants to host the nodes in a data centre, have basically the same simplicity, but the nodes can boot from a boot server that provides for DHCP, and also delivers the iPXE image to load, without the need for a USB stick in every Node.
+
+A boot server is not really necessary, but it helps! That server has a list of the MAC addresses of the nodes, and delivers the bootloader over PXE. The farmer is responsible to set-up the network, and configure the boot server.
+
+### Necessities
+
+The Farmer needs to:
+
+- Obtain an IPv6 prefix allocation from the provider. A `/64` will do, that is publicly reachable, but a `/48` is advisable if the farmer wants to provide IPv6 transit for User Networks
+- If IPv6 is not an option, obtain an IPv4 subnet from the provider. At least one IPv4 address per node is needed, where all IP addresses are publicly reachable.
+- Have the Nodes connected on that public network with a switch so that all Nodes are publicly reachable.
+- In case of multiple NICS, also make sure his farm is properly registered in BCDB, so that the Node's public IP Addresses are registered.
+- Properly list the MAC addresses of the Nodes, and configure the DHCP server to provide for an IP address, and in case of multiple NICs also provide for private IP addresses over DHCP per Node.
+- Make sure that after first boot, the Nodes are reachable.
+
+### IPv6
+
+IPv6, although already a real protocol since '98, has seen reluctant adoption over the time it exists. That mostly because ISPs and Carriers were reluctant to deploy it, and not seeing the need since the advent of NAT and private IP space, giving the false impression of security.
+But this month (10/2019), RIPE sent a mail to all it's LIRs that the last consecutive /22 in IPv4 has been allocated. Needless to say, but that makes the transition to IPv6 in 2019 of utmost importance and necessity.
+Hence, ZOS starts with IPv6, and IPv4 is merely an afterthought ;-)
+So in a nutshell: we greatly encourage Farmers to have IPv6 on the Node's network.
+
+### Routing/Firewalling
+
+Basically, the Nodes are self-protecting, in the sense that they provide no means at all to be accessed through listening processes at all. No service is active on the node itself, and User Networks function solely on an overlay.
+That also means that there is no need for a Farm admin to protect the Nodes from exterior access, albeit some DDoS protection might be a good idea.
+In the first phase we will still allow the Host OS (ZOS) to reply on ICMP ping requests, but that 'feature' might as well be blocked in the future, as once a Node is able to register itself, there is no real need to ever want to try to reach it.
+
+### Multi-NIC Nodes
+
+Nodes that Farmers deploy are typically multi-NIC Nodes, where one (typically a 1GBit NIC) can be used for getting a proper DHCP server running from where the Nodes can boot, and one other NIC (1Gbit or even 10GBit), that then is used for transfers of User Data, so that there is a clean separation, and possible injections bogus data is not possible.
+
+That means that there would be two networks, either by different physical switches, or by port-based VLANs in the switch (if there is only one).
+
+- Management NICs
+ The Management NIC will be used by ZOS to boot, and register itself to the GRID. Also, all communications from the Node to the Grid happens from there.
+- Public NICs
+
+### Farmers and the TFGrid
+
+A Node, being part of the Grid, has no concept of 'Farmer'. The only relationship for a Node with a Farmer is the fact that that is registered 'somewhere (TM)', and that a such workloads on a Node will be remunerated with Tokens. For the rest, a Node is a wholly stand-alone thing that participates in the Grid.
+
+```text
+ 172.16.1.0/24
+ 2a02:1807:1100:10::/64
++--------------------------------------+
+| +--------------+ | +-----------------------+
+| |Node ZOS | +-------+ | |
+| | +-------------+1GBit +--------------------+ 1GBit switch |
+| | | br-zos +-------+ | |
+| | | | | |
+| | | | | |
+| | | | +------------------+----+
+| +--------------+ | | +-----------+
+| | OOB Network | | |
+| | +----------+ ROUTER |
+| | | |
+| | | |
+| | | |
+| +------------+ | +----------+ |
+| | Public | | | | |
+| | container | | | +-----+-----+
+| | | | | |
+| | | | | |
+| +---+--------+ | +-------------------+--------+ |
+| | | | 10GBit Switch | |
+| br-pub| +-------+ | | |
+| +-----+10GBit +-------------------+ | +---------->
+| +-------+ | | Internet
+| | | |
+| | +----------------------------+
++--------------------------------------+
+ 185.69.167.128/26 Public network
+ 2a02:1807:1100:0::/64
+
+```
+
+Where the underlay part of the wireguard interfaces get instantiated in the Public container (namespace), and once created these wireguard interfaces get sent into the User Network (Network Resource), where a user can then configure the interface a he sees fit.
+
+The router of the farmer fulfills 2 roles:
+
+- NAT everything in the OOB network to the outside, so that nodes can start and register themselves, as well get tasks to execute from the BCDB.
+- Route the assigned IPv4 subnet and IPv6 public prefix on the public segment, to which the public container is connected.
+
+As such, in case that the farmer wants to provide IPv4 public access for grid proxies, the node will need at least one (1) IPv4 address. It's free to the farmer to assign IPv4 addresses to only a part of the Nodes.
+On the other hand, it is quite important to have a proper IPv6 setup, because things will work out better.
+
+It's the Farmer's task to set up the Router and the switches.
+
+In a simpler setup (small number of nodes for instance), the farmer could setup a single switch and make 2 port-based VLANs to separate OOB and Public, or even wit single-nic nodes, just put them directly on the public segment, but then he will have to provide a DHCP server on the Public network.
\ No newline at end of file
diff --git a/collections/documentation/farmers/advanced_networking/network_setup.md b/collections/documentation/farmers/advanced_networking/network_setup.md
new file mode 100644
index 0000000..1a0302d
--- /dev/null
+++ b/collections/documentation/farmers/advanced_networking/network_setup.md
@@ -0,0 +1,86 @@
+Network Setup
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Network Setup for Farmers](#network-setup-for-farmers)
+ - [Step 1. Testing for IPv6 Availability in Your Location](#step-1-testing-for-ipv6-availability-in-your-location)
+ - [Step 2. Choosing the Setup to Connect Your Nodes](#step-2-choosing-the-setup-to-connect-your-nodes)
+ - [2.1 Home Setup](#21-home-setup)
+ - [2.2 Data Center/Advanced Setup](#22-data-centeradvanced-setup)
+- [General Notes](#general-notes)
+
+***
+
+# Introduction
+
+0-OS nodes participating in the Threefold grid, need connectivity of course. They need to be able to communicate over
+the Internet with each-other in order to do various things:
+
+- download its OS modules
+- perform OS module upgrades
+- register itself to the grid, and send regular updates about it's status
+- query the grid for tasks to execute
+- build and run the Overlay Network
+- download flists and the effective files to cache
+
+The nodes themselves can have connectivity in a few different ways:
+
+- Only have RFC1918 private addresses, connected to the Internet through NAT, NO IPv6
+ Mostly, these are single-NIC (Network card) machines that can host some workloads through the Overlay Network, but
+ cant't expose services directly. These are HIDDEN nodes, and are mostly booted with an USB stick from
+ bootstrap.grid.tf .
+- Dual-stacked: having RFC1918 private IPv4 and public IPv6 , where the IPv6 addresses are received from a home router,
+but firewalled for outgoing traffic only. These nodes are effectively also HIDDEN
+- Nodes with 2 NICs, one that has effectively a NIC connected to a segment that has real public
+addresses (IPv4 and/or IPv6) and one NIC that is used for booting and local
+management. (OOB) (like in the drawing for farmer setup)
+
+For Farmers, we need to have Nodes to be reachable over IPv6, so that the nodes can:
+
+- expose services to be proxied into containers/vms
+- act as aggregating nodes for Overlay Networks for HIDDEN Nodes
+
+Some Nodes in Farms should also have a publicly reachable IPv4, to make sure that clients that only have IPv4 can
+effectively reach exposed services.
+
+But we need to stress the importance of IPv6 availability when you're running a multi-node farm in a datacentre: as the
+grid is boldly claiming to be a new Internet, we should make sure we adhere to the new protocols that are future-proof.
+Hence: IPv6 is the base, and IPv4 is just there to accomodate the transition.
+
+Nowadays, RIPE can't even hand out consecutive /22 IPv4 blocks any more for new LIRs, so you'll be bound to market to
+get IPv4, mostly at rates of 10-15 Euro per IP. Things tend to get costly that way.
+
+So anyway, IPv6 is not an afterthought in 0-OS, we're starting with it.
+
+# Network Setup for Farmers
+
+This is a quick manual to what is needed for connecting a node with zero-OS V2.0
+
+## Step 1. Testing for IPv6 Availability in Your Location
+As descibed above the network in which the node is instaleld has to be IPv6 enabled. This is not an afterthought as we are building a new internet it has to ba based on the new and forward looking IP addressing scheme. This is something you have to investigate, negotiate with you connectivity provider. Many (but not all home connectivity products and certainly most datacenters can provide you with IPv6. There are many sources of infromation on how to test and check whether your connection is IPv6 enabled, [here is a starting point](http://www.ipv6enabled.org/ipv6_enabled/ipv6_enable.php)
+
+## Step 2. Choosing the Setup to Connect Your Nodes
+
+Once you have established that you have IPv6 enabled on the network you are about to deploy, you have to make sure that there is an IPv6 DHCP facility available. Zero-OS does not work with static IPv6 addresses (at this point in time). So you have choose and create one of the following setups:
+
+### 2.1 Home Setup
+
+Use your (home) ISP router Ipv6 DHCP capabilities to provide (private) IPv6 addresses. The principle will work the same as for IPv4 home connections, everything happens enabled by Network Adress Translation (just like anything else that uses internet connectivity). This should be relatively straightforward if you have established that your conenction has IPv6 enabled.
+
+### 2.2 Data Center/Advanced Setup
+
+In this situation there are many options on how to setup you node. This requires you as the expert to make a few decisions on how to connect what what the best setup is that you can support for the operaitonal time of your farm. The same basics principles apply:
+ - You have to have a block of (public) IPv6 routed to you router, or you have to have your router setup to provide Network Address Translation (NAT)
+ - You have to have a DHCP server in your network that manages and controls IPV6 ip adress leases. Depending on your specific setup you have this DHCP server manage a public IPv6y range which makes all nodes directly connected to the public internet or you have this DHCP server manage a private block og IPv6 addresses which makes all you nodes connect to the internet through NAT.
+
+As a farmer you are in charge of selecting and creating the appropriate network setup for your farm.
+
+# General Notes
+
+The above setup will allows your node(s) to appear in explorer on the TFGrid and will allow you to earn farming tokens. At stated in the introduction ThreeFold is creating next generation internet capacity and therefore has IPv6 as it's base building block. Connecting to the current (dominant) IPv4 network happens for IT workloads through so called webgateways. As the word sais these are gateways that provide connectivity between the currenct leading IPv4 adressing scheme and IPv6.
+
+We have started a forum where people share their experiences and configurations. This will be work in progress and forever growing.
+
+**IMPORTANT**: You as a farmer do not need access to IPV4 to be able to rent capacity for IT workloads that need to be visible on IPV4, this is something that can happen elsewhere on the TFGrid.
+
\ No newline at end of file
diff --git a/collections/documentation/farmers/advanced_networking/networking_overview.md b/collections/documentation/farmers/advanced_networking/networking_overview.md
new file mode 100644
index 0000000..c4bc322
--- /dev/null
+++ b/collections/documentation/farmers/advanced_networking/networking_overview.md
@@ -0,0 +1,94 @@
+ Networking Overview
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Possible Configurations](#possible-configurations)
+- [Overall Requirements](#overall-requirements)
+- [Notes and Warnings](#notes-and-warnings)
+ - [Management Interfaces](#management-interfaces)
+ - [Data Center Cable Management](#data-center-cable-management)
+ - [Static IP Uplink](#static-ip-uplink)
+- [Testing the Setup](#testing-the-setup)
+- [Questions and Feedback](#questions-and-feedback)
+
+***
+
+## Introduction
+
+In this section, we provide advanced networking tips for farms with public IPs and in data centers (DC). The information available in this section is a combination of documentation from ThreeFold and tips and advice from community members who experienced first-hand the creation of ThreeFold farms that make use of public IPs block in data centers, personal data centers and home farms. A special thank you to those who contributed to improving the TFGrid and its knowledge base documentation.
+
+## Possible Configurations
+
+For farmers who have public IPs, extra considerations are needed in setting up the network of the farm. We will go through the main considerations in this section.
+
+First, we must acknowledge that by the open-source and design of ThreeFold farming, a farm can range from a simple [single 3Node](../3node_building/3node_building.md) setup, to a multi-rack farm hosted in a typical data center, and everything in-between, from the farmer experiencing with public IP blocks, to the entrepreneur who builds their own data center at home.
+
+There are thus many types of farms and each will have varying configurations. The simplest way to set up a farm has been extensively discussed in the first steps of creating a farm. But what are the other more complex configurations possible? Let's go through some of those:
+
+- Network link
+ - DC provides a network link into the farmer's rack
+- Router and switch
+ - The farmer provider their own router and switch
+ - DC provides a router and/or switch in the rack
+- Gateway IP and public IP
+ - Gateway IP provided is in the same range as the public IPs
+ - Gateway IP is in a different range than the public IPs
+- Segmenting
+ - Farmer segments the OOB ("Zos"/private) interfaces and the public interfaces into
+ - separate VLANs, OR;
+ - uses separate switches altogether
+ - No segmenting is actually necessary, farmer connects all interfaces to one switch
+
+## Overall Requirements
+
+There are overall requirements for any 3Node farm using IP address blocks in a data centere or at home:
+
+- There must be at least one interface that provide DHCP to each node
+- Public IPs must be routable from at least one interface
+
+Note that redundancy can help in avoiding single point of failure [(SPOF)](https://en.wikipedia.org/wiki/Single_point_of_failure).
+
+## Notes and Warnings
+
+### Management Interfaces
+
+You should make sure to never expose management interfaces to the public internet.
+
+
+### Data Center Cable Management
+
+It's important to have a good cable management, especially if you are in a data center. Proper cable management will improve the cooling streams of your farm. There shouldn't be any cable in front of the fans. This way, your servers will last longer. If you want to patch a rack, you have to have all lenght of patch cables from 30cm to 3m. Also, try to keep the cables as short as possible. Arrange the cables in bundles of eight and lead them to the sides of the rack as much as possible for optimal airflow.
+
+
+
+
+
+### Static IP Uplink
+
+If your DC uplink is established by simple static IP (which is the case in most DCs), there is a simple setup possible. Note that if you have to use PPPoE or pptp/L2TP (like a consumer internet connection at most homes), this would not work.
+
+If your WAN is established by static IP, you can simply attach the WAN uplink provided by the DC to one of the switches (and not to the WAN-side of your own router). Then, the WAN-side of the router needs to be attached to the switch too. By doing so, your nodes will be able to connect directly to the DC gateway, in the same way that the router is connecting its WAN-side to the gateway, without the public IP traffic being routed/bridged through the router (bypassing).
+
+With a network configured like this, it is absolutely not important on which ports you connect which NIC of your nodes. You can just randomly plug them anywhere. But there is one restriction: the DC uplink must use a static IP. Dynamic IP would also not work because you would then have two DHCP servers in the same physical network (the one from the DC and your own router).
+
+## Testing the Setup
+
+Manual and automatic validation of the network of a farm are possible. More information on automatic validation will be added in the future.
+
+You can test the network of your farm manually by deploying a workload on your 3Nodes with either a gateway or a public IP reserved.
+
+## Questions and Feedback
+
+If you have any questions, you can ask the ThreeFold community for help on the [ThreeFold Forum](http://forum.threefold.io/) or on the [ThreeFold Farmer Chat](https://t.me/threefoldfarmers) on Telegram.
\ No newline at end of file
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/3node_diy_desktop.md b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/3node_diy_desktop.md
new file mode 100644
index 0000000..de74eb0
--- /dev/null
+++ b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/3node_diy_desktop.md
@@ -0,0 +1,406 @@
+Building a DIY 3Node: Desktop Computer
+
+In the following 3Node DIY guide, you will learn how to turn a Dell Optiplex 7020 into a 3Node farming on the ThreeFold Grid.
+
+Note that the process is similar for other desktop computers.
+
+
+
+
+
+Table of Contents
+
+
+
+- [Prerequisite](#prerequisite)
+ - [DIY 3Node Computer Requirements](#diy-3node-computer-requirements)
+ - [DIY 3Node Material List](#diy-3node-material-list)
+- [1. Create a Farm](#1-create-a-farm)
+ - [Using Dashboard](#using-dashboard)
+ - [Using TF Connect App](#using-tf-connect-app)
+- [2. Create a Zero-OS Bootstrap Image](#2-create-a-zero-os-bootstrap-image)
+ - [Download the Zero-OS Boostrap Image](#download-the-zero-os-boostrap-image)
+ - [Burn the Zero-OS Bootstrap Image](#burn-the-zero-os-bootstrap-image)
+- [3. Set the Hardware](#3-set-the-hardware)
+- [4. Wipe All the Disks](#4-wipe-all-the-disks)
+ - [1. Create a Linux Boostrap Image](#1-create-a-linux-boostrap-image)
+ - [2. Boot Linux in Try Mode](#2-boot-linux-in-try-mode)
+ - [3. Use wipefs to Wipe All Disks](#3-use-wipefs-to-wipe-all-disks)
+- [5. Set the BIOS/UEFI](#5-set-the-biosuefi)
+ - [The Essential Features of BIOS/UEFI for a 3Node](#the-essential-features-of-biosuefi-for-a-3node)
+ - [Set the BIOS/UEFI on a Dell Optiplex 7020](#set-the-biosuefi-on-a-dell-optiplex-7020)
+- [6. Boot the 3Node](#6-boot-the-3node)
+ - [Check the Node Status](#check-the-node-status)
+ - [Farming Rewards Distribution](#farming-rewards-distribution)
+- [Additional Information](#additional-information)
+
+***
+
+
+
+
+
+# Prerequisite
+
+## DIY 3Node Computer Requirements
+
+
+
+Any computer with the following specifications can be used as a DIY 3Node.
+
+- Any 64-bit hardware with an Intel or AMD processor chip.
+- Servers, desktops and mini computers type hardware are compatible.
+- A minimum of 500 GB of SSD and a bare minimum of 2 GB of RAM is required.
+- A ratio of 100GB of SSD and 8GB of RAM per thread is recommended.
+- A wired ethernet connection is highly recommended to maximize reliability and the ability to farm TFT.
+- A [passmark](https://www.passmark.com/) of 1000 per core is recommended and will be a minimum requirement in the future.
+
+In this guide, we are using a Dell Optiplex 7020. It constitutes a perfect affordable entry DIY 3Node as it can be bought refurbished with the proper ratio of 100GB of SSD and 8GB of RAM per thread, and this, without any need of upgrades or updates.
+
+
+
+## DIY 3Node Material List
+
+
+
+* Any computer respecting the DIY 3Node Computer Requirements stated above
+* Ethernet cable
+* Router + Modem
+* Surge Protector
+* 2x USB key 4 Go
+* Android/iOS Phone
+* Computer monitor and cable, keyboard and mouse
+* MAC/Linux/Windows Computer
+
+
+
+
+
+# 1. Create a Farm
+
+You can create a farm with either the ThreeFold Dashboard or the ThreeFold Connect app.
+
+## Using Dashboard
+
+The Dashboard section contains all the information required to [create a farm](../../../dashboard/farms/your_farms.md).
+
+## Using TF Connect App
+
+You can [create a ThreeFold farm](../../../threefold_token/storing_tft/tf_connect_app.md) with the ThreeFold Connect App.
+
+
+# 2. Create a Zero-OS Bootstrap Image
+
+## Download the Zero-OS Boostrap Image
+
+We will now learn how to create a Zero-OS Bootstrap Image in order to boot a DIY 3Node.
+
+Go on the [ThreeFold Zero-OS Bootstrap Link](https://v3.bootstrap.grid.tf) as shown above.
+
+![Farming_Create_Farm_21](./img/farming_createfarm_21.png)
+
+This is the Zero-OS v3 Bootstrapping page.
+
+![Farming_Create_Farm_22](./img/farming_createfarm_22.png)
+
+Write your farm ID and choose production mode.
+
+![Farming_Create_Farm_23](./img/farming_createfarm_23.png)
+
+Download the bootstrap image. Next, we will burn the bootstrap image.
+
+
+
+
+
+## Burn the Zero-OS Bootstrap Image
+
+
+
+For **MAC**, **Linux** and **Windows**, you can use [BalenaEtcher](https://www.balena.io/etcher/) to load/flash the image on a USB stick. This program also formats the USB in the process. This will work for the option **EFI IMG** for UEFI boot, and with the option **USB** for BIOS boot. Simply follow the steps presented to you and make sure you select the correct bootstrap image file you downloaded previously.
+
+General Steps:
+
+1. Download BalenaEtcher at [https://balena.io/etcher](https://balena.io/etcher)
+
+![3node_diy_desktop_42.png](img/3node_diy_desktop_42.png)
+
+![3node_diy_desktop_43.png](img/3node_diy_desktop_43.png)
+
+![3node_diy_desktop_44.png](img/3node_diy_desktop_44.png)
+
+![3node_diy_desktop_45.png](img/3node_diy_desktop_45.png)
+
+![3node_diy_desktop_48.png](img/3node_diy_desktop_48.png)
+
+![3node_diy_desktop_49.png](img/3node_diy_desktop_49.png)
+
+2. Open BalenaEtcher
+
+![3node_diy_desktop_50.png](img/3node_diy_desktop_50.png)
+
+3. Select **Flash from file**
+
+![3node_diy_desktop_52.png](img/3node_diy_desktop_52.png)
+
+1. Find and select the bootstrap image in your computer
+
+2. Select **Target** (your USB key)
+
+![3node_diy_desktop_53.png](img/3node_diy_desktop_53.png)
+
+![3node_diy_desktop_54.png](img/3node_diy_desktop_54.png)
+
+6. Select **Flash**
+
+![3node_diy_desktop_55.png](img/3node_diy_desktop_55.png)
+
+![3node_diy_desktop_56.png](img/3node_diy_desktop_56.png)
+
+
+That's it. Now you have a bootstrap image on Zero-OS as a bootable removable media device.
+
+
+
+
+
+# 3. Set the Hardware
+
+Setting the hardware of this DIY 3node is very easy as there are no updates or upgrades needed. Simply unbox the computer and plug everything.
+
+![3node_diy_desktop_40.png](img/3node_diy_desktop_40.jpeg)
+
+![3node_diy_desktop_38.png](img/3node_diy_desktop_38.jpeg)
+
+![3node_diy_desktop_30.png](img/3node_diy_desktop_30.jpeg)
+
+![3node_diy_desktop_29.png](img/3node_diy_desktop_29.jpeg)
+
+Plug the computer cable in the surge protector
+
+![3node_diy_desktop_6.png](img/3node_diy_desktop_6.png)
+
+Connect the computer cable, the ethernet cable, the mouse and keyboard cable and the monitor cable.
+
+![3node_diy_desktop_13.png](img/3node_diy_desktop_13.jpeg)
+
+Plug the ethernet cable in the router (or the switch)
+
+![3node_diy_desktop_6.png](img/3node_diy_desktop_3.png)
+
+
+
+
+
+# 4. Wipe All the Disks
+
+In this section, we will learn how to create a Linux bootstrap image, boot it in Try mode and then wipe all the disks in your 3Node. To create a Linux boostrap image, follow the same process as when we burnt the Zero-OS Boostrap Image.
+
+
+
+## 1. Create a Linux Boostrap Image
+
+
+
+1. Download the Ubuntu 20.04 ISO file [here](https://releases.ubuntu.com/20.04/)
+2. Burn the ISO image on a USB key with Balena Etcher
+
+
+
+## 2. Boot Linux in Try Mode
+
+
+
+1. Insert your Linux boostrap image USB key in your computer and boot it
+2. During boot, press F12 to enter into Settings
+3. Select your booting device, here it is: *UEFI: USB DISK 2.0*
+
+![3node_diy_desktop_107.png](img/3node_diy_desktop_107.jpeg)
+
+4. Select Try or install Ubuntu
+
+![3node_diy_desktop_106.png](img/3node_diy_desktop_106.jpeg)
+
+5. Select Try Ubuntu
+
+![3node_diy_desktop_105.png](img/3node_diy_desktop_105.jpeg)
+
+
+
+## 3. Use wipefs to Wipe All Disks
+
+
+
+Once Ubuntu is booted, you will land on the main page.
+
+![3node_diy_desktop_67.png](img/3node_diy_desktop_67.png)
+
+At the bottom left of the screen, click on Applications.
+
+![3node_diy_desktop_68.png](img/3node_diy_desktop_68.png)
+
+In Applications, select Terminal.
+
+![3node_diy_desktop_69.png](img/3node_diy_desktop_69.png)
+
+If you don't see it, write terminal in the search box.
+
+![3node_diy_desktop_70.png](img/3node_diy_desktop_70.png)
+
+You will land in the Ubuntu Terminal.
+
+![3node_diy_desktop_71.png](img/3node_diy_desktop_71.png)
+
+Write the command line *lsblk* as shown below. You will then see the disks in your computer. You want to wipe the main disk, but not the USB key we are using, named *sdb* here. We can see here that the SSD disk, *sda*, has 3 partitions: *sda1*, *sda2*, *sda3*. Note that when wiping the disks, we want no partition.
+
+In this case, the disk we want to wipe is *sda*.
+
+![3node_diy_desktop_72.png](img/3node_diy_desktop_72.png)
+
+Write the command line *sudo wipefs -a /dev/sda*. This will wipe the disk *sda*.
+
+![3node_diy_desktop_73.png](img/3node_diy_desktop_73.png)
+
+If you write the command line *lsblk* once more, you should see that your SSD disk has no more partition. The disk has been properly wiped.
+
+![3node_diy_desktop_74.png](img/3node_diy_desktop_74.png)
+
+Power off the computer by selecting *Power Off* after having clicked on the button at the top right of the screen as shown below.
+
+![3node_diy_desktop_75.png](img/3node_diy_desktop_75.png)
+
+That's it! The disks are all wiped. All that is left now is to set the BIOS/UEFI settings and then boot the 3Node!
+
+
+
+
+
+# 5. Set the BIOS/UEFI
+
+Before booting the main operating system, in our case Zero-OS, a computer will boot in either BIOS or UEFI mode. Older systems use BIOS and newer systems uses UEFI. Both BIOS and UEFI are low-lewel softwares needed to interact between the hardware and the main OS of the computer. Note that BIOS is also called Legacy BIOS.
+
+## The Essential Features of BIOS/UEFI for a 3Node
+
+
+
+There are certain things that you should make sure are set properly on your 3Node.
+
+As a general advice, you can Load Defaults (Settings) on your BIOS, then make sure the options below are set properly.
+
+* Choose the correct combination of BIOS/UEFI and bootstrap image on [https://bootstrap.grid.tf/](https://bootstrap.grid.tf/)
+ * Newer system will use UEFI --> the Dell Optiplex 7020 uses UEFI
+ * Bootstrap image: *EFI IMG* and *EFI FILE*
+ * Older system will use Legacy BIOS
+ * Bootstrap image: *ISO* and *USB*
+* Set *Multi-Processor* and *Hyperthreading* at Enabled
+ * Sometimes, it will be written *Virtual Cores*, or *Logical Cores*.
+* Set *Virtualization* at Enabled
+ * On Intel, it is denoted as *CPU virtualization* and on ASUS, it is denoted as *SVM*.
+ * Make sure virtualization is enabled and look for the precise terms in your specific BIOS/UEFI.
+* Enable *Network Stack* (sometimes called *Network Boot*)
+* Set *AC Recovery* at *Last Power State*
+ * This will make sure your 3Node restarts after losing power momentarily.
+* Select the proper *Boot Sequence* for the 3Node to boot Zero-OS from your bootstrap image
+ * e.g., if you have a USB key as a bootstrap image, select it in *Boot Sequence*
+* Set *Server Lookup Method* (or the equivalent) at *DNS*.
+ * Only use Static IP if you know what you are doing.
+ * Your router will automatically assign a dynamic IP address to your 3Node when it connects to Internet.
+* Set *Client Address Method* (or the equivalent) at *DHCP*. Only use Static IP if you know what you are doing.
+ * Your router will automatically assign a dynamic IP address to your 3Node when it connects to Internet.
+* *Secure Boot* should be left at disabled
+ * Enable it if you know what you are doing. Otherwise, it can be set at disabled.
+
+
+
+
+
+## Set the BIOS/UEFI on a Dell Optiplex 7020
+
+
+
+1. Insert your Zero-OS boostrap image USB key in your computer and boot it.
+2. During boot, press F12 to enter into *Settings* then choose *BIOS Setup*.
+
+![3node_diy_desktop_104.jpeg](img/3node_diy_desktop_109.png)
+
+3. In BIOS Setup, click on Load Default and confirm by clicking on *OK*
+
+![3node_diy_desktop_115.png](img/3node_diy_desktop_115.png)
+
+4. Leave the BIOS Setup (Exit) and re-enter. This will set the default settings.
+
+5. Go through each page and make sure you are following the guidelines set in the section Essential Features as shown in the following pictures.
+
+![3node_diy_desktop_.png](img/3node_diy_desktop_116.png)
+
+![3node_diy_desktop_.png](img/3node_diy_desktop_117.png)
+
+![3node_diy_desktop_.png](img/3node_diy_desktop_118.png)
+
+![3node_diy_desktop_.png](img/3node_diy_desktop_114.png)
+
+![3node_diy_desktop_.png](img/3node_diy_desktop_127.png)
+
+![3node_diy_desktop_.png](img/3node_diy_desktop_120.png)
+
+![3node_diy_desktop_.png](img/3node_diy_desktop_128.png)
+
+![3node_diy_desktop_.png](img/3node_diy_desktop_122.png)
+
+![3node_diy_desktop_.png](img/3node_diy_desktop_123.png)
+
+![3node_diy_desktop_.png](img/3node_diy_desktop_129.png)
+
+![3node_diy_desktop_.png](img/3node_diy_desktop_125.png)
+
+
+6. Once you are done, click on *Exit* and then click *Yes* to save your changes. The 3node will now boot Zero-OS.
+
+![3node_diy_desktop_126.png](img/3node_diy_desktop_126.png)
+
+
+
+
+
+# 6. Boot the 3Node
+
+If your BIOS/UEFI settings are set properly and you have the Zero-OS bootstrap image USB key plugged in, your 3node should automatically boot Zero-OS every time that it is turned on.
+
+1. Power on the 3Node with the Zero-OS boostrap image USB key
+2. Let the 3Node load Zero-OS
+ * The first time it boots, the 3node will register to the TF Grid
+3. Verify the 3Node's status on ThreeFold Explorer
+
+The first time you boot a 3Node, it will be written: “This node is not registered (farmer : NameOfFarm). This is normal. The Grid will create a node ID and you will be able to see it on screen. This can take a couple of minutes.
+
+This is the final screen you should see when your 3Node is connected to the ThreeFold Grid. Note that it is normal if it is written *no public config* next to *PUB* as we did not set any public IP address.
+
+Naturally, your node ID as well as your farm ID and name will be shown.
+
+![3node_diy_desktop_76.png](img/3node_diy_desktop_130.png)
+
+Once you have your node ID, you can also go on the ThreeFold Dashboard to see your 3Node and verify that your 3Node is online.
+
+
+
+
+
+## Check the Node Status
+
+You can use the [Node Finder](../../../dashboard/deploy/node_finder.md) on the [TF Dashboard](https://dashboard.grid.tf/) to verify that the node is online.
+
+Enter your node ID and click **Apply**.
+
+## Farming Rewards Distribution
+
+
+
+The farming reward will be sent once per month directly in your ThreeFold Connect App wallet. Farming rewards are usually distributed around the 5th of each month.
+
+
+
+# Additional Information
+
+Congratulations, you now have built your first ThreeFold 3Node server!
+
+If you have any questions, you can ask the ThreeFold community for help on the [ThreeFold Forum](https://forum.threefold.io/) or on the [ThreeFold Telegram Farmer Group](https://t.me/threefoldfarmers).
\ No newline at end of file
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_104.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_104.jpeg
new file mode 100644
index 0000000..80e850c
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_104.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_105.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_105.jpeg
new file mode 100644
index 0000000..088454a
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_105.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_106.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_106.jpeg
new file mode 100644
index 0000000..a24ff11
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_106.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_107.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_107.jpeg
new file mode 100644
index 0000000..8ab6fdb
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_107.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_108.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_108.png
new file mode 100644
index 0000000..b426984
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_108.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_109.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_109.png
new file mode 100644
index 0000000..45841c5
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_109.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_110.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_110.png
new file mode 100644
index 0000000..71377f1
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_110.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_111.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_111.png
new file mode 100644
index 0000000..7b5eeb4
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_111.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_112.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_112.png
new file mode 100644
index 0000000..ef5fcb3
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_112.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_114.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_114.png
new file mode 100644
index 0000000..8ce20bd
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_114.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_115.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_115.png
new file mode 100644
index 0000000..ca2260e
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_115.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_116.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_116.png
new file mode 100644
index 0000000..65cf981
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_116.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_117.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_117.png
new file mode 100644
index 0000000..8c67944
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_117.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_118.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_118.png
new file mode 100644
index 0000000..726b43c
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_118.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_119.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_119.png
new file mode 100644
index 0000000..2dff9b7
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_119.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_120.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_120.png
new file mode 100644
index 0000000..507141d
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_120.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_121.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_121.png
new file mode 100644
index 0000000..53775ac
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_121.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_122.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_122.png
new file mode 100644
index 0000000..8a2fbab
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_122.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_123.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_123.png
new file mode 100644
index 0000000..77c28e2
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_123.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_124.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_124.png
new file mode 100644
index 0000000..a4b7660
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_124.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_125.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_125.png
new file mode 100644
index 0000000..d23c25a
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_125.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_126.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_126.png
new file mode 100644
index 0000000..c35857e
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_126.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_127.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_127.png
new file mode 100644
index 0000000..1a0249c
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_127.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_128.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_128.png
new file mode 100644
index 0000000..2e0c8c7
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_128.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_129.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_129.png
new file mode 100644
index 0000000..734aa84
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_129.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_13.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_13.jpeg
new file mode 100644
index 0000000..94ffab0
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_13.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_130.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_130.png
new file mode 100644
index 0000000..0630b55
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_130.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_23.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_23.jpeg
new file mode 100644
index 0000000..b4a953a
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_23.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_25.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_25.jpeg
new file mode 100644
index 0000000..66fa069
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_25.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_26.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_26.jpeg
new file mode 100644
index 0000000..f773e8f
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_26.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_27.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_27.jpeg
new file mode 100644
index 0000000..e8eaf53
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_27.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_28.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_28.jpeg
new file mode 100644
index 0000000..8333b2b
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_28.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_29.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_29.jpeg
new file mode 100644
index 0000000..b86b9af
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_29.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_3.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_3.jpeg
new file mode 100644
index 0000000..30456be
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_3.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_3.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_3.png
new file mode 100644
index 0000000..7442f5d
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_3.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_30.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_30.jpeg
new file mode 100644
index 0000000..622d1f9
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_30.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_31.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_31.jpeg
new file mode 100644
index 0000000..2ffe2de
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_31.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_38.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_38.jpeg
new file mode 100644
index 0000000..a818dd6
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_38.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_40.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_40.jpeg
new file mode 100644
index 0000000..3093e7e
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_40.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_42.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_42.png
new file mode 100644
index 0000000..e7336f6
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_42.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_43.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_43.png
new file mode 100644
index 0000000..1c139d8
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_43.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_44.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_44.png
new file mode 100644
index 0000000..dae09b3
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_44.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_45.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_45.png
new file mode 100644
index 0000000..668e6eb
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_45.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_48.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_48.png
new file mode 100644
index 0000000..b789d93
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_48.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_49.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_49.png
new file mode 100644
index 0000000..057fbc9
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_49.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_50.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_50.png
new file mode 100644
index 0000000..17b0329
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_50.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_52.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_52.png
new file mode 100644
index 0000000..b19d799
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_52.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_53.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_53.png
new file mode 100644
index 0000000..600c1ec
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_53.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_54.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_54.png
new file mode 100644
index 0000000..56a5237
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_54.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_55.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_55.png
new file mode 100644
index 0000000..81da31e
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_55.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_56.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_56.png
new file mode 100644
index 0000000..a2590e7
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_56.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_6.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_6.jpeg
new file mode 100644
index 0000000..8bd003e
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_6.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_6.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_6.png
new file mode 100644
index 0000000..d893e54
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_6.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_67.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_67.png
new file mode 100644
index 0000000..46ae916
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_67.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_68.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_68.png
new file mode 100644
index 0000000..0842c19
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_68.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_69.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_69.png
new file mode 100644
index 0000000..2d2b6b9
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_69.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_70.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_70.png
new file mode 100644
index 0000000..192a5d7
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_70.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_71.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_71.png
new file mode 100644
index 0000000..7609bb2
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_71.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_72.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_72.png
new file mode 100644
index 0000000..14efc8b
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_72.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_73.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_73.png
new file mode 100644
index 0000000..d645b73
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_73.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_74.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_74.png
new file mode 100644
index 0000000..439fdb7
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_74.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_75.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_75.png
new file mode 100644
index 0000000..7d7c8d2
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_75.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_76.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_76.jpeg
new file mode 100644
index 0000000..84cf1b0
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_76.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_76.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_76.png
new file mode 100644
index 0000000..263ddf7
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_76.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_77.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_77.jpeg
new file mode 100644
index 0000000..25ed402
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_77.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_78.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_78.jpeg
new file mode 100644
index 0000000..0c61dc3
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_78.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_80.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_80.jpeg
new file mode 100644
index 0000000..9658fb5
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_80.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_81.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_81.jpeg
new file mode 100644
index 0000000..9e69a2c
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_81.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_83.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_83.jpeg
new file mode 100644
index 0000000..b98ed0b
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_83.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_86.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_86.jpeg
new file mode 100644
index 0000000..c3a77b6
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_86.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_88.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_88.jpeg
new file mode 100644
index 0000000..0ed0169
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_88.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_89.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_89.jpeg
new file mode 100644
index 0000000..8711387
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_89.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_90.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_90.jpeg
new file mode 100644
index 0000000..41e9d80
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_90.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_91.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_91.jpeg
new file mode 100644
index 0000000..836ffae
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_91.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_92.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_92.jpeg
new file mode 100644
index 0000000..a671268
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_92.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_94.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_94.jpeg
new file mode 100644
index 0000000..1cf820e
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_94.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_95.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_95.jpeg
new file mode 100644
index 0000000..0e8fcb2
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_95.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_96.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_96.jpeg
new file mode 100644
index 0000000..71f9646
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_96.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_97.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_97.jpeg
new file mode 100644
index 0000000..782eaea
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_97.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_98.jpeg b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_98.jpeg
new file mode 100644
index 0000000..187342c
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/3node_diy_desktop_98.jpeg differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_001.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_001.png
new file mode 100644
index 0000000..b666fce
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_001.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_002.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_002.png
new file mode 100644
index 0000000..415a4b3
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_002.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_003.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_003.png
new file mode 100644
index 0000000..47d3e2f
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_003.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_004.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_004.png
new file mode 100644
index 0000000..0e1aa75
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_004.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_005.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_005.png
new file mode 100644
index 0000000..6acfa59
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_005.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_006.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_006.png
new file mode 100644
index 0000000..f874fcf
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_006.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_25.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_25.png
new file mode 100644
index 0000000..d85e2bf
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_25.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_26.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_26.png
new file mode 100644
index 0000000..3a22175
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_26.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_27.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_27.png
new file mode 100644
index 0000000..dd04e45
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_27.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_28.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_28.png
new file mode 100644
index 0000000..36f2eda
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_28.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_29.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_29.png
new file mode 100644
index 0000000..e7d6d40
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_29.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_30.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_30.png
new file mode 100644
index 0000000..d810d13
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_30.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_1.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_1.png
new file mode 100644
index 0000000..95545d8
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_1.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_10.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_10.png
new file mode 100644
index 0000000..deec259
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_10.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_11.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_11.png
new file mode 100644
index 0000000..ceb60ae
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_11.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_12.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_12.png
new file mode 100644
index 0000000..70e8053
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_12.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_13.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_13.png
new file mode 100644
index 0000000..1016cb8
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_13.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_14.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_14.png
new file mode 100644
index 0000000..7468258
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_14.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_15.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_15.png
new file mode 100644
index 0000000..0a34bd0
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_15.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_16.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_16.png
new file mode 100644
index 0000000..1e32275
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_16.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_17.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_17.png
new file mode 100644
index 0000000..0ee82ea
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_17.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_18.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_18.png
new file mode 100644
index 0000000..b3f4386
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_18.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_19.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_19.png
new file mode 100644
index 0000000..74d7963
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_19.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_2.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_2.png
new file mode 100644
index 0000000..488f970
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_2.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_20.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_20.png
new file mode 100644
index 0000000..05a80e8
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_20.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_21.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_21.png
new file mode 100644
index 0000000..ab36d45
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_21.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_22.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_22.png
new file mode 100644
index 0000000..7cfc6ba
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_22.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_23.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_23.png
new file mode 100644
index 0000000..3537771
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_23.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_24.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_24.png
new file mode 100644
index 0000000..4b2da96
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_24.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_3.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_3.png
new file mode 100644
index 0000000..349ec17
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_3.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_4.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_4.png
new file mode 100644
index 0000000..b717bfe
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_4.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_5.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_5.png
new file mode 100644
index 0000000..a022a1f
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_5.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_6.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_6.png
new file mode 100644
index 0000000..922b77b
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_6.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_7.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_7.png
new file mode 100644
index 0000000..e4bdb6e
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_7.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_8.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_8.png
new file mode 100644
index 0000000..e8d023c
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_8.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_9.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_9.png
new file mode 100644
index 0000000..b6e852a
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_createfarm_9.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_1.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_1.png
new file mode 100644
index 0000000..f1b4b1e
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_1.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_10.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_10.png
new file mode 100644
index 0000000..e8c547b
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_10.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_11.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_11.png
new file mode 100644
index 0000000..d0e58bc
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_11.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_12.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_12.png
new file mode 100644
index 0000000..980e33f
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_12.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_13.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_13.png
new file mode 100644
index 0000000..c4956c5
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_13.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_14.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_14.png
new file mode 100644
index 0000000..f251968
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_14.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_15.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_15.png
new file mode 100644
index 0000000..efd806e
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_15.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_16.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_16.png
new file mode 100644
index 0000000..72d1994
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_16.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_17.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_17.png
new file mode 100644
index 0000000..cc9ba7f
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_17.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_18.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_18.png
new file mode 100644
index 0000000..8fe5d2f
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_18.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_19.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_19.png
new file mode 100644
index 0000000..6c6796c
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_19.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_2.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_2.png
new file mode 100644
index 0000000..eccb743
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_2.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_20.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_20.png
new file mode 100644
index 0000000..da5c0ca
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_20.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_21.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_21.png
new file mode 100644
index 0000000..27abf0c
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_21.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_22.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_22.png
new file mode 100644
index 0000000..adde6b0
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_22.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_23.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_23.png
new file mode 100644
index 0000000..aabf0a0
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_23.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_24.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_24.png
new file mode 100644
index 0000000..9f09996
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_24.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_25.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_25.png
new file mode 100644
index 0000000..c37a89c
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_25.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_26.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_26.png
new file mode 100644
index 0000000..5cb2f75
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_26.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_27.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_27.png
new file mode 100644
index 0000000..c790ae0
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_27.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_28.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_28.png
new file mode 100644
index 0000000..2648b65
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_28.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_29.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_29.png
new file mode 100644
index 0000000..52a66cd
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_29.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_3.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_3.png
new file mode 100644
index 0000000..f0b992c
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_3.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_30.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_30.png
new file mode 100644
index 0000000..2e30308
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_30.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_31.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_31.png
new file mode 100644
index 0000000..ff1b9ac
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_31.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_32.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_32.png
new file mode 100644
index 0000000..ab617bf
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_32.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_33.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_33.png
new file mode 100644
index 0000000..3c6d591
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_33.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_34.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_34.png
new file mode 100644
index 0000000..944688f
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_34.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_35.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_35.png
new file mode 100644
index 0000000..845bd35
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_35.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_36.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_36.png
new file mode 100644
index 0000000..b294bcb
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_36.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_37.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_37.png
new file mode 100644
index 0000000..c61e06e
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_37.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_38.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_38.png
new file mode 100644
index 0000000..732f98a
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_38.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_39.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_39.png
new file mode 100644
index 0000000..0c78bc0
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_39.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_4.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_4.png
new file mode 100644
index 0000000..5047a5c
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_4.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_40.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_40.png
new file mode 100644
index 0000000..6651627
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_40.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_41.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_41.png
new file mode 100644
index 0000000..839e929
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_41.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_42.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_42.png
new file mode 100644
index 0000000..5f84480
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_42.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_43.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_43.png
new file mode 100644
index 0000000..ba1e751
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_43.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_44.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_44.png
new file mode 100644
index 0000000..4a10071
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_44.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_45.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_45.png
new file mode 100644
index 0000000..0d6e86d
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_45.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_46.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_46.png
new file mode 100644
index 0000000..2039941
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_46.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_47.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_47.png
new file mode 100644
index 0000000..71ea72e
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_47.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_48.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_48.png
new file mode 100644
index 0000000..fd357c7
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_48.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_49.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_49.png
new file mode 100644
index 0000000..0f82c97
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_49.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_5.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_5.png
new file mode 100644
index 0000000..6c960d9
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_5.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_50.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_50.png
new file mode 100644
index 0000000..0ecbc06
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_50.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_51.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_51.png
new file mode 100644
index 0000000..86ecf04
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_51.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_52.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_52.png
new file mode 100644
index 0000000..9d2b5a4
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_52.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_53.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_53.png
new file mode 100644
index 0000000..141c24d
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_53.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_54.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_54.png
new file mode 100644
index 0000000..2c97d1b
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_54.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_55.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_55.png
new file mode 100644
index 0000000..2d7fe0e
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_55.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_56.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_56.png
new file mode 100644
index 0000000..e090e08
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_56.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_57.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_57.png
new file mode 100644
index 0000000..b9289a4
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_57.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_58.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_58.png
new file mode 100644
index 0000000..a78d6d8
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_58.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_59.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_59.png
new file mode 100644
index 0000000..4f6e7f2
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_59.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_6.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_6.png
new file mode 100644
index 0000000..1ccab0c
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_6.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_7.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_7.png
new file mode 100644
index 0000000..be8eea7
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_7.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_8.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_8.png
new file mode 100644
index 0000000..1239f37
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_8.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_9.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_9.png
new file mode 100644
index 0000000..4a6acad
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_tf_wallet_9.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_wallet_10.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_wallet_10.png
new file mode 100644
index 0000000..7ae95fc
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_wallet_10.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_wallet_11.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_wallet_11.png
new file mode 100644
index 0000000..6ce3313
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_wallet_11.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_wallet_12.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_wallet_12.png
new file mode 100644
index 0000000..b26559e
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_wallet_12.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_wallet_13.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_wallet_13.png
new file mode 100644
index 0000000..9486799
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_wallet_13.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_wallet_5.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_wallet_5.png
new file mode 100644
index 0000000..e5fc90b
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_wallet_5.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_wallet_7.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_wallet_7.png
new file mode 100644
index 0000000..ce1f1ba
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_wallet_7.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_wallet_8.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_wallet_8.png
new file mode 100644
index 0000000..b8039d9
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_wallet_8.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_wallet_9.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_wallet_9.png
new file mode 100644
index 0000000..882bcf7
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/farming_wallet_9.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/old_backup/farming_createfarm_12.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/old_backup/farming_createfarm_12.png
new file mode 100644
index 0000000..0bcdb35
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/old_backup/farming_createfarm_12.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/old_backup/farming_createfarm_13.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/old_backup/farming_createfarm_13.png
new file mode 100644
index 0000000..477500e
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/old_backup/farming_createfarm_13.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/readme.md b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/readme.md
new file mode 100644
index 0000000..6f45a2d
--- /dev/null
+++ b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/img/readme.md
@@ -0,0 +1 @@
+# Images of Farming Guide documentation for Threefold Manual 3.0
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/readme.md b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/readme.md
new file mode 100644
index 0000000..362d32f
--- /dev/null
+++ b/collections/documentation/farmers/complete_diy_guides/3node_diy_desktop/readme.md
@@ -0,0 +1,19 @@
+# DIY 3node Desktop for the Threefold Manual 3.0
+* The **diy_3node_desktop.md** file
+ * contains the easiest DIY 3node guide possible
+ * no upgrades
+ * no updates
+
+* The **diy_3node_desktop.pdf** file
+ * can be used as an offline reference
+ * can be shared among the Threefold community
+
+Updates: This DIY Guide will be turned into a short 1 minute video to share.
+
+
+# Threefold Ebooks
+
+1. FAQ
+2. Complete Farming Guide
+3. DIY 3node Desktop Computer - Farming Guide
+4. DIY 3node Rack Server - Farming Guide
\ No newline at end of file
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/3node_diy_rack_server.md b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/3node_diy_rack_server.md
new file mode 100644
index 0000000..3686915
--- /dev/null
+++ b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/3node_diy_rack_server.md
@@ -0,0 +1,373 @@
+Building a DIY 3Node: Rack Server
+
+In the following 3Node DIY guide, you will learn how to turn a Dell server (R620, R720) into a 3Node farming on the ThreeFold Grid.
+
+Note that the process is similar for other rack servers.
+
+
+Table of Contents
+
+
+
+- [Setting Up the Hardware](#setting-up-the-hardware)
+ - [Avoiding Static Discharge](#avoiding-static-discharge)
+ - [Setting the M.2 NVME SSD Disk with the PCIe Adaptor](#setting-the-m2-nvme-ssd-disk-with-the-pcie-adaptor)
+ - [Checking the RAM sticks](#checking-the-ram-sticks)
+ - [General Rules when Installing RAM Sticks](#general-rules-when-installing-ram-sticks)
+ - [Procedure to Install RAM Sticks](#procedure-to-install-ram-sticks)
+ - [Installing the SSD Disks](#installing-the-ssd-disks)
+ - [Plugging the 3node Server](#plugging-the-3node-server)
+ - [Removing the DVD Optical Drive - Installing a SSD disk in the DVD Optical Drive Slot](#removing-the-dvd-optical-drive---installing-a-ssd-disk-in-the-dvd-optical-drive-slot)
+ - [Using Onboard Storage - RAID Controller Details](#using-onboard-storage---raid-controller-details)
+- [Zero-OS Bootstrap Image](#zero-os-bootstrap-image)
+ - [Creating a Farm](#creating-a-farm)
+ - [Using Dashboard](#using-dashboard)
+ - [Using TF Connect App](#using-tf-connect-app)
+ - [Wiping All the Disks](#wiping-all-the-disks)
+ - [Downloading the Zero-OS Bootstrap Image](#downloading-the-zero-os-bootstrap-image)
+ - [DVD ISO BIOS Image](#dvd-iso-bios-image)
+ - [USB BIOS Image](#usb-bios-image)
+- [BIOS Settings](#bios-settings)
+ - [Processor Settings](#processor-settings)
+ - [Boot Settings](#boot-settings)
+- [Booting the 3Node](#booting-the-3node)
+- [Additional Information](#additional-information)
+ - [Differences between the R620 and the R720](#differences-between-the-r620-and-the-r720)
+ - [Different CPUs and RAMs Configurations for 3Node Dell Servers](#different-cpus-and-rams-configurations-for-3node-dell-servers)
+- [Closing Words](#closing-words)
+
+***
+
+# Setting Up the Hardware
+
+![3node_diy_rack_server_1](./img/3node_diy_rack_server_1.png)
+
+Dell R620 1U server
+
+
+## Avoiding Static Discharge
+
+![3node_diy_rack_server_2](./img/3node_diy_rack_server_2.png)
+
+
+Some will recommend to wear anti-static gloves as shown here. If you don’t have anti-static gloves, remember this:
+
+> Always touch the metal side of the server before manipulating the hardware.
+
+Your hands will discharge the static on the outside of the box, which is secure.
+
+## Setting the M.2 NVME SSD Disk with the PCIe Adaptor
+
+![3node_diy_rack_server_3](./img/3node_diy_rack_server_3.png)
+
+Here is one of the two 2TB SSD NVME m.2 that we will install on the server. Above the SSD is the PCIe Gen 3, x4 that we will use to connect the SSD to the server.
+
+![3node_diy_rack_server_4](./img/3node_diy_rack_server_4.png)
+
+You can see at the left of the adaptor that there is a metal piece that can be used to hold more firmly the PCIe adaptor and the SSD. We will remove it for this DIY build. Why? Because it is not necessary as the adaptor can hold the weight of the SSD. Also, this metal piece is full while the brackets in the server have holes in it. This will ensure a better airflow and thus less heat.
+
+![3node_diy_rack_server_5](./img/3node_diy_rack_server_5.png)
+
+We remove the screws with a star screwdriver.
+
+![3node_diy_rack_server_6](./img/3node_diy_rack_server_6.png)
+
+![3node_diy_rack_server_7](./img/3node_diy_rack_server_7.png)
+
+This SSD already has a heatsink. There is no need to use the heatsink included in the PCIe adaptor kit. If you remove the heatsink or the sticker under the SSD, you will lose your 5 years warranty.
+
+![3node_diy_rack_server_8](./img/3node_diy_rack_server_8.png)
+
+When you put the SSD in the adaptor, make sure you have the opening in line with the adaptor.
+
+![3node_diy_rack_server_9](./img/3node_diy_rack_server_9.png)
+
+![3node_diy_rack_server_10](./img/3node_diy_rack_server_10.png)
+
+Fitting in the SSD takes some force. Do not overdo it and take your time!
+
+![3node_diy_rack_server_11](./img/3node_diy_rack_server_11.png)
+
+It’s normal that the unscrewed part is lifting in the air before you screw the SSD on the adaptor.
+
+![3node_diy_rack_server_12](./img/3node_diy_rack_server_12.png)
+
+To screw the SSD in place, use the screwdriver included in the PCIe adaptor kit.
+
+![3node_diy_rack_server_13](./img/3node_diy_rack_server_13.png)
+
+![3node_diy_rack_server_14](./img/3node_diy_rack_server_14.png)
+
+Now that’s a steady SSD!
+
+## Checking the RAM sticks
+
+![3node_diy_rack_server_15](./img/3node_diy_rack_server_15.png)
+
+It’s now time to get under the hood! Make sure the case is at the unlocked position. If you need to turn it to unlocked position, use a flathead screwdriver or a similar tool.
+
+![3node_diy_rack_server_16](./img/3node_diy_rack_server_16.png)
+
+![3node_diy_rack_server_17](./img/3node_diy_rack_server_17.png)
+
+Lift up the lock and the top server plate should glide to the back. You can remove the top of the server.
+
+![3node_diy_rack_server_18](./img/3node_diy_rack_server_18.png)
+
+Here’s the full story! R620 and all!
+
+![3node_diy_rack_server_19](./img/3node_diy_rack_server_19.png)
+
+![3node_diy_rack_server_20](./img/3node_diy_rack_server_20.png)
+
+![3node_diy_rack_server_21](./img/3node_diy_rack_server_21.png)
+
+To remove this plastic piece, simply lift with your fingers at the designated spot (follow the blue line!).
+
+![3node_diy_rack_server_22](./img/3node_diy_rack_server_22.png)
+
+Here’s the RAMs! This R620 came already equipped with 256GB of rams dispersed in 16x16GB sticks. If you need to add the RAM sticks yourself, make sure you are doing it correctly. The FAQ covers some basic information for RAM installation.
+
+![3node_diy_rack_server_23](./img/3node_diy_rack_server_23.png)
+
+To remove a stick, push on the clips on both sides. You can do it one at a time if you want. Make sure it doesn’t pop out and fall on a hardware piece! Once the clips are opened, pull out the RAM stick by holding it on the sides. This will ensure it does not get damaged.
+
+![3node_diy_rack_server_24](./img/3node_diy_rack_server_24.png)
+
+Here’s the RAM in it’s purest form!
+
+![3node_diy_rack_server_25](./img/3node_diy_rack_server_25.png)
+
+Here you can see that the gap is not in the middle of the RAM stick. You must be careful when inserting the RAM. Make sure you have the gap aligned with the RAM holder.
+
+![3node_diy_rack_server_26](./img/3node_diy_rack_server_26.png)
+
+When you want to put a RAM stick in its slot, make sure the plastic holders on the sides are opened and insert the RAM stick. Make sure you align the RAM stick properly. You can then push on one side at a time until the RAM stick clicks in. You can do it both sides at once if you are at ease.
+
+### General Rules when Installing RAM Sticks
+
+First, always use RAM sticks of the same size and type. It should be noted on your motherboard which slots to populate first.
+
+As a general guide, there is usually 2 slots A and B, with each 2 memory stick entries. You must then install the ram sticks on A1 and B1 in order to achieve dual channel, then A2 and B2 if you have more (visual order: A1 A2 B1 B2).
+
+#### Procedure to Install RAM Sticks
+
+You want to start with your largest sticks, evenly distributed between both processors and work your way down to your smallest.
+
+As an example, let's say you have 2 processors and 4x 16GB sticks and 4x 8GB sticks. The arrangement would be A1-16GB, B1-16GB, A2-16GB, B2-16GB, A3-8GB, B3-8GB, A4-8GB, B4-8GB.
+
+Avoid odd numbers as well. You optimally want pairs. So if you only have 5x 8GB sticks, only install 4 until you have an even 6.
+
+
+## Installing the SSD Disks
+
+
+![3node_diy_rack_server_27](./img/3node_diy_rack_server_27.png)
+
+To put back the plastic protector, simply align the plastic piece with the two nudges in the metal case.
+
+![3node_diy_rack_server_28](./img/3node_diy_rack_server_28.png)
+
+![3node_diy_rack_server_29](./img/3node_diy_rack_server_29.png)
+
+We will now remove this PCIe riser in order to connect the SSDs.
+
+![3node_diy_rack_server_30](./img/3node_diy_rack_server_30.png)
+
+Optional step: put the SSDs and the PCIe riser next to each other so they can talk and break the ice. They will get to learn one another before going into the server to farm TFT.
+
+![3node_diy_rack_server_31](./img/3node_diy_rack_server_31.png)
+
+Just like with RAM sticks, you want to make sure you are aligned with the slot.
+
+![3node_diy_rack_server_32](./img/3node_diy_rack_server_32.png)
+
+Next, push the adaptor inside the riser’s opening. This takes some force too. If you are well aligned, it should be done with ease.
+
+![3node_diy_rack_server_33](./img/3node_diy_rack_server_33.png)
+
+This is what the riser looks like with the two SSDs installed. Now you simply need to put the riser back inside the server.
+
+![3node_diy_rack_server_34](./img/3node_diy_rack_server_34.png)
+
+Push down on the riser to insert it properly.
+
+![3node_diy_rack_server_35](./img/3node_diy_rack_server_35.png)
+
+It’s good to notice that the inside of the top plate of the server has great pictures showing how to manipulate the hardware.
+
+
+
+## Plugging the 3node Server
+
+
+
+![3node_diy_rack_server_36](./img/3node_diy_rack_server_36.png)
+
+Now you will want to plug in the power cable in the PSU. Here we show two 495W PSUs. With 256GB of RAM and two SSDs NVME, it is better to use two 750W PSUs. Note that this server will only use around 100W at idle. There are two power cables for redundancy. The unit does not need more than one to function.
+
+On a 15A/120V breaker, you can have more than one server. But note that, at full load, this server can use up to 400W. In this case, no more than 3 servers could be plugged on the same breaker. Make sure you adapt to your current situation (server's power consumption, electric breaker, etc.).
+
+![3node_diy_rack_server_37](./img/3node_diy_rack_server_37.png)
+
+Plugging in the power cable is pretty straightforward. Just make sure you have the 3 pins oriented properly!
+
+![3node_diy_rack_server_38](./img/3node_diy_rack_server_38.png)
+
+It is highly recommended to plug the power cable in a surge protector. If you have unsteady electricity at your location, it might be good to use a UPS, uninterrupted power supply. A surge protector is essential to avoid overpowering and damaging the server.
+
+![3node_diy_rack_server_39](./img/3node_diy_rack_server_39.png)
+
+![3node_diy_rack_server_40](./img/3node_diy_rack_server_40.png)
+
+![3node_diy_rack_server_41](./img/3node_diy_rack_server_41.png)
+
+Before starting the server, you can plug in the monitor and the keyboard as well as the ethernet cable. Make sure you plug the ethernet cable in one of the four NIC ports.
+
+![3node_diy_rack_server_42](./img/3node_diy_rack_server_42.png)
+
+Now, power it on!
+
+![3node_diy_rack_server_43](./img/3node_diy_rack_server_43.png)
+
+The server is booting.
+
+
+## Removing the DVD Optical Drive - Installing a SSD disk in the DVD Optical Drive Slot
+
+
+![3node_diy_rack_server_44](./img/3node_diy_rack_server_44.png)
+
+![3node_diy_rack_server_45](./img/3node_diy_rack_server_45.png)
+
+If you want to change the DVD optical drive, push where indicated and remove the power and SATA cables.
+
+It is possible to install a SSD disk in there. To do so, use a SATA HDD hard drive caddy CD/DVD **9.5mm** and put in a SATA III 2.5" disk. The caddy is not necessary. You could simply remove the standard CD/DVD caddy and plug the SATA disk.
+
+The hardware part is done. Next, you will want to set the BIOS properly as well as get the bootstrap image of Zero-OS. Before we get into this, let's have some information on using the onboard storage of your 3node server.
+
+
+## Using Onboard Storage - RAID Controller Details
+
+
+If you want to use the onboard storage on your server, you will probably need to flash the RAID card or do some adjustment in order for Zero-OS to recognize your disks.
+
+You can use the onboard storage on a server without RAID. You can [re-flash](https://fohdeesha.com/docs/perc.html) the RAID card, turn on HBA/non-RAID mode, or install a different card. It's usually easy to set servers such as a HP Proliant with the HBA mode.
+
+For Dell servers, you can either cross-flash the RAID controller with an “IT-mode-Firmware” (see this [video](https://www.youtube.com/watch?v=h5nb09VksYw)) or get a DELL H310-controller (which has the non-RAID option). Otherwise, as shown in this guide, you can install a NVME SSD with a PCIe adaptor, and turn off the RAID controller.
+
+Note that for Dell R610 and R710, you can re-flash the RAID card. For the R910, you can’t re-flash the card. In this case, you will need to get a LSI Dell card.
+
+# Zero-OS Bootstrap Image
+
+With R620 and R720 Dell servers, UEFI does not work well. You will want to use either a DVD or a USB in BIOS mode.
+
+Go on https://bootstrap.grid.tf/ and download the appropriate image: option **ISO** for the DVD and option **USB** for BIOS USB (not UEFI).
+
+Write your farmer ID and make sure you select production mode.
+
+## Creating a Farm
+
+You can create a farm with either the ThreeFold Dashboard or the ThreeFold Connect app.
+
+### Using Dashboard
+
+The Dashboard section contains all the information required to [create a farm](../../../dashboard/farms/your_farms.md).
+
+### Using TF Connect App
+
+You can [create a ThreeFold farm](../../../threefold_token/storing_tft/tf_connect_app.md) with the ThreeFold Connect App.
+
+## Wiping All the Disks
+
+You might need to wipe your disks if they are not brand new. To wipe your disks, read the section [Wipe All the Disks](../../3node_building/4_wipe_all_disks.md) of the ThreeFold Farming Documentation.
+
+## Downloading the Zero-OS Bootstrap Image
+
+You can then download the [Zero-OS bootstrap image](https://v3.bootstrap.grid.tf) for your farm.
+
+![3node_diy_rack_server_46](./img/3node_diy_rack_server_46.png)
+
+![3node_diy_rack_server_47](./img/3node_diy_rack_server_47.png)
+
+Use the ISO image for DVD boot and the USB image for USB BIOS boot (not UEFI). We use the farm ID 1 here as an example. Put your own farm ID.
+
+### DVD ISO BIOS Image
+For the ISO image, download the file and burn it on a DVD.
+
+### USB BIOS Image
+Note: the USB key must be formatted before burning the Zero-OS bootstrap image.
+
+For Windows, MAC and Linux, you can use [balenaEtcher](https://www.balena.io/etcher/), a free and open source software that will let you create a bootstrap image on a USB key, while also formatting the USB key at the same time.
+
+This is the **easiest way** to burn your Zero-OS bootstrap image. All the steps are clearly explained within the software.
+
+For Windows, you can also use Rufus.
+
+For the USB image, with Linux, you can also go through the command window and write:
+
+> dd status=progress if=FILELOCATION.ISO(or .IMG) of=/dev/sd*.
+
+Here the * is to indicate that you must adjust according to your disk. To see your disks, write **lsblk** in the command window. Make sure you select the proper disk.
+
+
+# BIOS Settings
+
+Before starting the server, plug in the USB bootstrap image. You can also insert the DVD once the server is on.
+
+When you start the server, press F2 to get into System Setup.
+
+Then, select System BIOS. In System BIOS settings, select Processor Settings.
+
+Note: More details are available for BIOS Settings in this [documentation](../../3node_building/5_set_bios_uefi.md).
+
+## Processor Settings
+
+Make sure you have enabled the Logical Processor (Hyper Threading with HP). This turns 8 cores into 16 virtual cores. You can set QPI Speed at Maximum data rate. Make sure you set All to Number of Cores per Processor. You can adjust the Processor Core speed and Processor Bus Speed for specific uses.
+
+It is also good to take a look at the Processors and make sure the hardware is correct.
+
+## Boot Settings
+
+Go to System BIOS Settings and select Boot Settings. In Boot Settings, choose BIOS and not UEFI as the Boot Mode. You need to save your preferences and comeback to select BIOS Boot Settings.
+
+Once back in BIOS Boot Settings, go to Boot Sequence. Depending on your bootstrap image of Zero-OS, select either the USB key or the Optical Drive CD-DVD option. The name of the USB key can be Drive C or else depending on where you plugged it and your server model.
+
+You can also disable the booting options that are not need. It can be good to have a DVD and a USB key with the bootstrap images for redundancy. If one boot fails, the computer would try with the other options of the boot sequence. This can be done with 2 USB keys too.
+
+Boot Sequence Retry enabled will simply redo the booting sequence if the last time did not work.
+
+
+That's it. You've set the BIOS settings properly and now is time to boot the 3Node and connect to the ThreeFold Grid.
+
+You can then save your preferences and exit. Your server should restart and load the bootstrap image.
+
+# Booting the 3Node
+
+Once you've set the BIOS settings and restarted your computer, it will download the Zero-OS bootstrap image. This takes a couple of minutes.
+
+The first time you boot a 3Node, it will be written: “This node is not registered (farmer : NameOfFarm). This is normal. The Grid will create a node ID and you will be able to see it on screen. This can take a couple of minutes.
+
+Once you have your node ID, you can also go on the ThreeFold Explorer to see your 3Node and verify that the connection is recognized by the Explorer.
+
+# Additional Information
+## Differences between the R620 and the R720
+
+Note that the main difference between the R620 and the R720 is that the former is a 1U and the latter a 2U. 2U servers are usually less noisy and generate less heat than 1U servers since they have a greater volume. In the R720, fans are bigger and thus less noisy. This can be an important factor to consider. Both offer great performances and work well with Zero-OS.
+
+## Different CPUs and RAMs Configurations for 3Node Dell Servers
+
+Different CPUs and RAMs configurations are possible for the Dell R620/R720 servers.
+
+For example, you could replace the E5-2640 v2 CPUs for the E5-2695 V2. This would give you 48 Threads. You could then go with 12x32GB DDR3 LRDIMM. You would also need 5TB SSD total instead to get the proper ratio, which is 100GB of SSD and 8GB of RAM per virtual core (also called thread or logical core).
+
+Note that you cannot have more than 16 sticks of ECC DIMM on the R620/R720. For more sticks, you need LRDIMM as stated above.
+
+# Closing Words
+That's it. You have now built a DIY 3Node and you are farming on the ThreeFold Grid.
+
+If you encounter errors, you can read the section [Troubleshooting and Error Messages](../../../faq/faq.md#troubleshooting-and-error-messages) of the Farmer FAQ.
+
+If you have any questions, you can ask the ThreeFold community for help on the [ThreeFold Forum](https://forum.threefold.io/) or on the [ThreeFold Telegram Farmer Group](https://t.me/threefoldfarmers).
+
+>Welcome to the New Internet!
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_1.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_1.png
new file mode 100644
index 0000000..be135fb
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_1.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_10.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_10.png
new file mode 100644
index 0000000..59fd49b
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_10.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_11.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_11.png
new file mode 100644
index 0000000..e83f1a7
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_11.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_12.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_12.png
new file mode 100644
index 0000000..e74e59f
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_12.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_13.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_13.png
new file mode 100644
index 0000000..0aeb0e5
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_13.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_14.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_14.png
new file mode 100644
index 0000000..79b3798
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_14.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_15.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_15.png
new file mode 100644
index 0000000..50f9191
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_15.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_16.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_16.png
new file mode 100644
index 0000000..5f2c328
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_16.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_17.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_17.png
new file mode 100644
index 0000000..92cef03
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_17.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_18.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_18.png
new file mode 100644
index 0000000..a78b3f0
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_18.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_19.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_19.png
new file mode 100644
index 0000000..78212d9
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_19.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_2.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_2.png
new file mode 100644
index 0000000..d459bf3
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_2.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_20.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_20.png
new file mode 100644
index 0000000..5cdb630
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_20.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_21.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_21.png
new file mode 100644
index 0000000..84bfe28
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_21.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_22.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_22.png
new file mode 100644
index 0000000..62e7ee4
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_22.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_23.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_23.png
new file mode 100644
index 0000000..bdd4063
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_23.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_24.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_24.png
new file mode 100644
index 0000000..c9363fd
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_24.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_25.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_25.png
new file mode 100644
index 0000000..54cf093
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_25.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_26.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_26.png
new file mode 100644
index 0000000..3c4ee6d
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_26.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_27.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_27.png
new file mode 100644
index 0000000..3af18ad
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_27.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_28.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_28.png
new file mode 100644
index 0000000..07f7558
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_28.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_29.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_29.png
new file mode 100644
index 0000000..1f610e8
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_29.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_3.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_3.png
new file mode 100644
index 0000000..aa7d30d
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_3.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_30.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_30.png
new file mode 100644
index 0000000..965f8bf
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_30.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_31.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_31.png
new file mode 100644
index 0000000..b72cfd2
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_31.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_32.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_32.png
new file mode 100644
index 0000000..143efca
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_32.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_33.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_33.png
new file mode 100644
index 0000000..705e98a
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_33.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_34.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_34.png
new file mode 100644
index 0000000..47a2799
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_34.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_35.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_35.png
new file mode 100644
index 0000000..68448ab
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_35.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_36.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_36.png
new file mode 100644
index 0000000..20324e9
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_36.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_37.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_37.png
new file mode 100644
index 0000000..88f7691
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_37.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_38.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_38.png
new file mode 100644
index 0000000..b5210c5
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_38.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_39.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_39.png
new file mode 100644
index 0000000..6804b91
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_39.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_4.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_4.png
new file mode 100644
index 0000000..d5ae149
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_4.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_40.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_40.png
new file mode 100644
index 0000000..671b240
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_40.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_41.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_41.png
new file mode 100644
index 0000000..c090dc7
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_41.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_42.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_42.png
new file mode 100644
index 0000000..04909fb
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_42.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_43.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_43.png
new file mode 100644
index 0000000..5248051
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_43.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_44.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_44.png
new file mode 100644
index 0000000..a6f1d64
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_44.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_45.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_45.png
new file mode 100644
index 0000000..1dedf8c
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_45.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_46.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_46.png
new file mode 100644
index 0000000..b0c1140
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_46.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_47.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_47.png
new file mode 100644
index 0000000..9da0919
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_47.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_5.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_5.png
new file mode 100644
index 0000000..3b9f8d5
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_5.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_6.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_6.png
new file mode 100644
index 0000000..d7da48a
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_6.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_7.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_7.png
new file mode 100644
index 0000000..c464eae
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_7.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_8.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_8.png
new file mode 100644
index 0000000..e9afd38
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_8.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_9.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_9.png
new file mode 100644
index 0000000..0527f58
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/3node_diy_rack_server_9.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/farming_30.png b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/farming_30.png
new file mode 100644
index 0000000..d810d13
Binary files /dev/null and b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/farming_30.png differ
diff --git a/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/readme.md b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/readme.md
new file mode 100644
index 0000000..b525533
--- /dev/null
+++ b/collections/documentation/farmers/complete_diy_guides/3node_diy_rack_server/img/readme.md
@@ -0,0 +1 @@
+Image folder for /wethreepedia/3node_diy_rack_server
diff --git a/collections/documentation/farmers/complete_diy_guides/complete_diy_guides_readme.md b/collections/documentation/farmers/complete_diy_guides/complete_diy_guides_readme.md
new file mode 100644
index 0000000..c3b2360
--- /dev/null
+++ b/collections/documentation/farmers/complete_diy_guides/complete_diy_guides_readme.md
@@ -0,0 +1,10 @@
+ Complete DIY 3Node Guides
+
+This section of the ThreeFold Farmers book presents two short guides detailing how to build a DIY 3Node.
+
+A perfect start for newcomers is the Desktop guide. If you want to build a bigger 3Node, the Rack Server guide maybe the best fit your you!
+
+ Table of Contents
+
+- [3Node Desktop DIY Guide](./3node_diy_desktop/3node_diy_desktop.html)
+- [3Node Rack Server DIY Guide](./3node_diy_rack_server/3node_diy_rack_server.html)
\ No newline at end of file
diff --git a/collections/documentation/farmers/farmerbot/farmerbot_information.md b/collections/documentation/farmers/farmerbot/farmerbot_information.md
new file mode 100644
index 0000000..ef86e19
--- /dev/null
+++ b/collections/documentation/farmers/farmerbot/farmerbot_information.md
@@ -0,0 +1,436 @@
+
+ Farmerbot Additional Information
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Additional Information](#additional-information)
+ - [General Considerations](#general-considerations)
+ - [YAML Configuration File Template](#yaml-configuration-file-template)
+ - [Supported Commands and Flags](#supported-commands-and-flags)
+ - [Minimum specs to run the Farmerbot](#minimum-specs-to-run-the-farmerbot)
+ - [How to Prepare Your Farm for the Farmerbot with WOL](#how-to-prepare-your-farm-for-the-farmerbot-with-wol)
+ - [WOL Requirements](#wol-requirements)
+ - [Enabling WOL in the BIOS](#enabling-wol-in-the-bios)
+ - [ZOS Nodes and NIC](#zos-nodes-and-nic)
+ - [NIC Firmware and WOL](#nic-firmware-and-wol)
+ - [How to Move Your Farm to a Different Network](#how-to-move-your-farm-to-a-different-network)
+ - [The differences between power "state" and power "target"](#the-differences-between-power-state-and-power-target)
+ - [The differences between uptime, status and power state](#the-differences-between-uptime-status-and-power-state)
+ - [The sequence of events for a node managed by the Farmerbot](#the-sequence-of-events-for-a-node-managed-by-the-farmerbot)
+ - [The problematic states of a 3node set with the Farmerbot](#the-problematic-states-of-a-3node-set-with-the-farmerbot)
+ - [Using the ThreeFold Node Status Bot](#using-the-threefold-node-status-bot)
+ - [CPU overprovisioning](#cpu-overprovisioning)
+ - [Seed phrase and HEX secret](#seed-phrase-and-hex-secret)
+ - [Farmerbot directory tree](#farmerbot-directory-tree)
+ - [Dedicated Nodes and the Farmerbot](#dedicated-nodes-and-the-farmerbot)
+ - [Periodic wakeup](#periodic-wakeup)
+ - [Time period between random wakeups and power target update](#time-period-between-random-wakeups-and-power-target-update)
+ - [Upgrade to the new Farmerbot](#upgrade-to-the-new-farmerbot)
+ - [Set the Farmerbot without the mnemonics of a ThreeFold Dashboard account](#set-the-farmerbot-without-the-mnemonics-of-a-threefold-dashboard-account)
+- [Maintenance](#maintenance)
+ - [See the power state and power target of 3Nodes](#see-the-power-state-and-power-target-of-3nodes)
+ - [With GraphQL](#with-graphql)
+ - [With Grid Proxy](#with-grid-proxy)
+ - [Change manually the power target of a 3Node](#change-manually-the-power-target-of-a-3node)
+ - [Properly reboot the node if power target "Down" doesn't work](#properly-reboot-the-node-if-power-target-down-doesnt-work)
+ - [Add a 3Node to a running Farmerbot](#add-a-3node-to-a-running-farmerbot)
+ - [Update the Farmerbot with a new release](#update-the-farmerbot-with-a-new-release)
+- [Questions and Feedback](#questions-and-feedback)
+
+***
+
+# Introduction
+
+We present some general information concerning the Farmerbot as well as some advice for proper maintenance and troubleshooting.
+
+# Additional Information
+
+We present additional information to complement the [Quick Guide](farmerbot_quick.md).
+
+## General Considerations
+
+The Farmerbot doesn’t have to run physically in the farm since it instructs nodes over RMB to power on and off. The Farmerbot should be running at all time.
+
+The Farmerbot uses the nodes in the farm to send WOL packets to the node that needs to wakeup. For this reason, you need at least one node per farm to be powered on at all time. If you do not specify one node to be always on, the Farmerbot will randomly choose a node to stay on for each cycle. If all nodes in a subnet are powered off, there is no way other nodes in other subnets will be able to power them on again.
+
+Note that if you run the Farmerbot on your farm, it is logical to set the node running the Farmerbot as always on. In this case, it will always be this node that wakes up the other nodes.
+
+Currently, you can run only one Farmerbot per farm. Since you can only deploy one Farmerbot per farm, the Farmerbot can only run on one node at a time.
+
+Since you need at least one node to power up a second node, you can't use the Farmerbot with just one node. You need at least two 3Nodes in your farm to correctly use the Farmerbot.
+
+The Farmerbot gets its data completely from TFChain. This means that, unlike the previous version, the Farmerbot will not start all the nodes when it restarts.
+
+## YAML Configuration File Template
+
+The quick guide showed a simple form of the YAML configuration file. Here are all the parameters that can be set for the configuration file.
+
+```
+farm_id: ""
+included_nodes: [optional, if no nodes are added then the farmerbot will include all nodes in the farm, farm should contain at least 2 nodes]
+ - ""
+excluded_nodes:
+ - ""
+never_shutdown_nodes:
+ - ""
+power:
+ periodic_wake_up_start: ""
+ wake_up_threshold: ""
+ periodic_wake_up_limit: ""
+ overprovision_cpu: ""
+```
+
+## Supported Commands and Flags
+
+We present the different commands for the Farmerbot.
+
+- `start`: to start (power on) a node
+
+```bash
+farmerbot start --node -m -n dev -d
+```
+
+Where:
+
+```bash
+Flags:
+ --node uint32 the node ID you want to use
+
+Global Flags:
+-d, --debug by setting this flag the farmerbot will print debug logs too
+-m, --mnemonic string the mnemonic of the account of the farmer
+-n, --network string the grid network to use (default "main")
+-s, --seed string the hex seed of the account of the farmer
+-k, --key-type string key type for mnemonic (default "sr25519")
+```
+
+- `start all`: to start (power on) all nodes in a farm
+
+```bash
+farmerbot start all --farm -m -n dev -d
+```
+
+Where:
+
+```bash
+Flags:
+ --farm uint32 the farm ID you want to start your nodes ins
+
+Global Flags:
+-d, --debug by setting this flag the farmerbot will print debug logs too
+-m, --mnemonic string the mnemonic of the account of the farmer
+-n, --network string the grid network to use (default "main")
+-s, --seed string the hex seed of the account of the farmer
+-k, --key-type string key type for mnemonic (default "sr25519")
+```
+
+- `version`: to get the current version of farmerbot
+
+```bash
+farmerbot version
+```
+
+## Minimum specs to run the Farmerbot
+
+The Farmerbot can run on any computer/server, it could even run on a laptop, so to speak. As long as it has an internet connection, the Farmerbot will be working fine.
+
+The Farmerbot runs fine on a VM with a single vcore and 2GB of RAM. For the storage, you need to have room for Docker and its dependencies. Thus 1 or 2GB of free storage, with the OS already installed, should be sufficient.
+
+## How to Prepare Your Farm for the Farmerbot with WOL
+
+ZOS can utilize 2 NIC's (Network Interface Card) of a node (server, workstation, desktop, ..). The first NIC on the motherboard will always be what we call the ZOS/dmz NIC, the second one is used for public config's (Gateway, public IP's for workloads, ..). So if you don't have public IP's in your farm, only the first NIC of your ZOS node will be used. This subnet is where the farmerbot operates. If you do have public IP's the same applies.
+
+Wake On LAN (WOL) is used to be able to boot (start) a ZOS node remotely that was shut down by the farmerbot. It works by sending what is called a 'magic packet' to the NIC MAC address of a ZOS node. If that NIC is setup correctly, aka 'listening' for the packet, the node will start up, post and boot ZOS. The farmerbot will keep a list of MAC addresses for nodes under it's management, so it knows where to send the packet if it's required.
+
+## WOL Requirements
+
+WOL comes with a few requirements. We list them in the sections that follow.
+
+### Enabling WOL in the BIOS
+
+Enable WOL in the BIOS of your ZOS node.
+
+A ZOS node must be capable of doing WOL. Have a look at your node hardware / BIOS manual. If so make sure to enable it in the BIOS! A bit of research will quickly tell you how to enable for your hardware. Some older motherboards do not support this, sometimes you can be lucky it does after a BIOS upgrade, but that is brand/model specific.
+
+Some examples:
+
+![farmerbot_bios_1|517x291](img/farmerbot_bios_1.jpeg)
+
+![farmerbot_bios_2|499x375](img/farmerbot_bios_2.jpeg)
+
+### ZOS Nodes and NIC
+
+All your ZOS nodes and their first NIC (ZOS/dmz) should be in the same network subnet (also called network segment or broadcast domain).
+
+This requires some basic network knowledge. WOL packets can not be send across different subnets by default, it can but this requires specific configuration on the firewall that connects the two subnets. Though cross-subnet WOL is currently not supported by the farmerbot.
+
+A 'magic' WOL packet is sent only on networking layer 2 (L2 or the 'data link layer') based on MAC address. So not on L3 based on ip address. This is why all nodes that should be brought up via WOL, need to be in the same subnet.
+
+You can check if this is the case like this: if for example one node has the ip 192.168.0.20/24, then all other nodes should have an ip between 192.168.0.1 and 192.168.0.254. You can calculate subnet ranges easely here: https://www.tunnelsup.com/subnet-calculator/
+
+So for the 192.168.0.0/24 example, you can see the range under 'Usable Host Range':
+
+![farmerbot_bios_3|499x500](img/farmerbot_bios_3.png)
+
+### NIC Firmware and WOL
+
+Some NIC's require WOL to be set on the NIC firmware.
+
+This is fully handled by ZOS. Every time ZOS boots it will enable WOL on links if they require it. So if a ZOS node then is added to a farmerbot, it will have WOL enabled on its NIC when it's turned off (by the farmerbot).
+
+Your farmerbot can be run on any system, including on a node. It doesn't have to be on the same network subnet as the nodes from the farm. The nodes of the farm on the other hand have to be in the same LAN. Don't hesitate to ask your technical questions here, we and the community will help you set things up!
+
+## How to Move Your Farm to a Different Network
+
+Note that the Farmerbot is currently available for Dev Net, QA Net, Test Net and Main Net. Thus, it might not be necessary to move your farm to a different network.
+
+To move your farm to a different network, you need to create a new bootstrap image for the new network instead of your current network. You should also wipe your 3Nodes' disks before moving to a different network.
+
+To download the Zero-OS bootstrap image, go to the usual bootstrap link [https://v3.bootstrap.grid.tf/](https://v3.bootstrap.grid.tf/) and select the network you want.
+
+![test_net|690x422](img/farmerbot_5.png)
+
+Once you have your new bootstrap image for the new network, [wipe your disks](../3node_building/4_wipe_all_disks.md), insert the new bootstrap image and reboot the 3Node.
+
+## The differences between power "state" and power "target"
+
+The target is what is set by the Farmerbot or can be set by the farmer manually on TF Chain. Power state can only be set by the node itself, in response to power targets it observes on chain.
+
+
+## The differences between uptime, status and power state
+
+There are three distinctly named endpoints or fields that exist in the back end systems:
+
+* Uptime
+ * number of seconds the node was up, as of it's last uptime report. This is the same on GraphQL and Grid Proxy.
+* Status
+ * this is a field that only exists on the Grid Proxy, which corresponds to whether the node sent an uptime report within the last 40 minutes.
+* Power state
+ * this is a field that only exists on GraphQL, and it's the self reported power state of the node. This only goes to "down" if the node shut itself down at request of the Farmerbot.
+
+## The sequence of events for a node managed by the Farmerbot
+
+The sequence of events for a node managed by farmerbot should look like this:
+
+1. Node is online. Target, state, and status are all "Up".
+2. Farmerbot sets node's target to "Down".
+3. Node sets its state to "Down" and then shuts off.
+4. Three hours later the status switches to "Down" because the node hasn't been updated.
+5. At periodic wake up time, Farmerbot sets node's target to "Up".
+6. Node receives WoL packet and starts booting.
+7. After boot is complete, node sets its state to "Up" and also submits uptime report.
+8. updatedAt is updated with time that uptime report was received and status changes to "Up".
+
+At that point the cycle is completed and will repeat.
+
+## The problematic states of a 3node set with the Farmerbot
+
+These are problematic states:
+
+1. Target is set to "Up" but state and status are "Down" for longer than normal boot time (node isn't responding).
+2. Target has been set to "Down" for longer than ~23.5 hours (farmerbot isn't working properly).
+3. Target is "Down" but state and status are up (Zos is potentially not responding to power target correctly).
+4. State is "Up" but status is "Down" (node shutdown unexpectedly).
+
+## Using the ThreeFold Node Status Bot
+
+You can use the [ThreeFold Node Status Bot](https://t.me/tfnodestatusbot) to see the nodes' status in relation to the Farmerbot.
+
+## CPU overprovisioning
+
+In the context of the ThreeFold grid, overprovisioning a CPU means that you can allocate more than one deployment to one CPU.
+
+In relation to the Farmerbot, you can set a value between 1 and 4 of how much the CPU can be overprovisioned. For example, a value of 2 means that the Farmerbot can allocate up to 2 deployments to one CPU.
+
+## Seed phrase and HEX secret
+
+When setting up the Farmerbot, you will need to enter either the seed phrase or the HEX secret of your farm. For farms created in the TF Connect app, the HEX secret from the app is correct. For farms created in the TF Dashboard, you'll need the seed phrase provided when you created the account.
+
+## Farmerbot directory tree
+
+As a general template, the directory tree of the Farmerbot will look like this:
+
+```
+└── farmerbot_directory
+ ├── .env
+ └── conf.yml
+```
+
+## Dedicated Nodes and the Farmerbot
+
+Dedicated nodes are managed like any other node. Nodes marked as dedicated can only be rented completely. Whenever a user wants to rent a dedicated node the user sends a find_node job to the farmerbot. The farmerbot will find such a node, power it on if it is down and reserve the full node (for 30 minutes). The user can then proceed with creating a rent contract for that node. The farmerbot will get that information and keep that node powered on. It will no longer return that node as a possible node in future find_node jobs. Whenever the rent contract is canceled the farmerbot will notice this and shutdown the node if the resource usage allows it.
+
+## Periodic wakeup
+
+The minimum period between two nodes to be waken up is currently 5 minutes. This means that every 5 minutes a new node wakes up during the periodic wakeup.
+
+Once all nodes are awaken, they all shut down at the same time, except the node that stays awaken to wake up the other during the next periodic wake.
+
+## Time period between random wakeups and power target update
+
+The time period between a random wakeup and the moment the power target is set to down is between 30 minutes and one hour.
+
+Whenever a random wakeup is initiated, the Farmerbot will wait 30 minutes for the node to be up. Once the node is up, the Farmerbot will keep that node up for 30 minutes for the two following reasons:
+
+* The node can send uptime report
+* If the node was put online for a given user deployment, this time priod gives ample time for the user to deploy their workload.
+
+This ensures an optimal user experience and reliablity in 3Nodes' reports.
+
+Note that each node managed by the Farmerbot will randomly wakeup on average 10 times a month.
+
+## Upgrade to the new Farmerbot
+
+If you are still running the old version of the Farmerbot (written in V), you can easily upgrade to the new Farmerbot (written in Go). You simply need to properly stop the old Farmerbot and then follow the new [Farmerbot guide](./farmerbot_quick.md).
+
+Here are the steps to properly stop the old Farmerbot.
+
+* Go to the diretory with the old Farmerbot docker files and fully stop the old Farmerbot:
+ ```
+ docker compose rm -f -s -v
+ ```
+* You should also make sure that there are no containers left from the previous runs. First, list all containers:
+ ```
+ docker container ls --all
+ ```
+* Then delete the remaining containers:
+ ```
+ docker container rm -f -v NAME_OF_CONTAINER
+ ```
+
+Once the old Farmerbot is properly stopped and deleted, follow the new [Farmerbot guide](./farmerbot_quick.md).
+
+## Set the Farmerbot without the mnemonics of a ThreeFold Dashboard account
+
+If you've lost the mnemonics associated with an account created on the ThreeFold Dashboard, it is still possible to set the Farmerbot with this account, but it's easier to simply create a new account and a new farm. Hopefully, the process is simple.
+
+- Create a new account on the Dashboard. This will generate a new twin.
+- Create a new farm and create new bootstrap images of your new farm.
+- Reboot your nodes with the new bootstrap images. This will automatically migrate your nodes with their current node IDs to the new farm.
+
+If you are using the Farmerbot, at this point, you will be able to set it with the mnemonics associated with the new farm.
+
+# Maintenance
+
+## See the power state and power target of 3Nodes
+
+### With GraphQL
+
+You can use [GraphQL](https://graphql.grid.tf/graphql) to see the power state and power target of 3Nodes.
+
+To find all nodes within one farm, use the following line with the proper farm ID (here we set farm **1** as an example):
+
+```
+query MyQuery {
+ nodes(where: {farmID_eq: 1}) {
+ power {
+ target
+ state
+ }
+ nodeID
+ }
+}
+```
+
+To find a specific node, write the following with the proper nodeID (here we set node **655** as an example):
+
+```
+query MyQuery {
+ nodes(where: {nodeID_eq: 655}) {
+ power {
+ state
+ }
+ nodeID
+ }
+}
+```
+
+### With Grid Proxy
+
+You can also see the power state and power target of a 3Node with Grid proxy.
+
+Use the following URL while adjusting the proper node ID (here we set node **1** as an example):
+
+```
+https://gridproxy.grid.tf/nodes/1
+```
+
+Then, in the response, you will see the following:
+
+```
+"power": {
+"state": "string",
+"target": "string"
+},
+```
+
+If the state and target are not defined, the string will be empty.
+
+## Change manually the power target of a 3Node
+
+You can use the Polkadot Extrinsics for this.
+
+* Go to the Polkadot.js.org website's endpoint based on the network of your 3Node:
+ * [Main net](https://polkadot.js.org/apps/?rpc=wss%3A%2F%2Ftfchain.grid.tf#/extrinsics)
+ * [Test net](https://polkadot.js.org/apps/?rpc=wss%3A%2F%2Ftfchain.test.grid.tf#/extrinsics)
+ * [Dev net](https://polkadot.js.org/apps/?rpc=wss%3A%2F%2Ftfchain.dev.grid.tf#/extrinsics)
+ * [QA net](https://polkadot.js.org/apps/?rpc=wss%3A%2F%2Ftfchain.qa.grid.tf#/extrinsics)
+* Make sure that **Developer -> Extrinsics** is selected
+* Select your account
+* Select **tfgridModule**
+* Select **changepowertarget(nodeId,powerTarget)**
+* Select the node you want to change the power target
+* Select the power target (**Up** or **Down**)
+* Click **Submit Transaction** at the bottom of the page
+
+
+
+## Properly reboot the node if power target "Down" doesn't work
+
+* Set the power target to "Down" manually
+* Reboot the node and wait for it to set its power state to "Down"
+* Once power target and state are both set to "Down", you can manually power off the node and reboot it
+
+## Add a 3Node to a running Farmerbot
+
+If the Farmerbot is running and you want to add a new 3Node to your farm, you can proceed as follows.
+
+- Boot the new 3Node
+ - Once the node is registered to the grid, a new node ID will be generated
+- If you set the section `included_nodes` in the YAML configuration file
+ - Add the new node ID to the configuration file
+- Restart the Farmerbot with the systemd command `restart` (in this example, the service is called `farmerbot`)
+ ```
+ systemctl restart farmerbot
+ ```
+
+## Update the Farmerbot with a new release
+
+There are only a few steps needed to update the Farmerbot to a new release.
+
+- Download the latest [ThreeFold tfgrid-sdk-go release](https://github.com/threefoldtech/tfgrid-sdk-go/releases) and extract the farmerbot for your specific setup (here we use `x86_64`). On the line `wget ...`, make sure to replace `` with the latest Farmerbot release.
+ ```
+ wget https://github.com/threefoldtech/tfgrid-sdk-go/releases/download//tfgrid-sdk-go_Linux_x86_64.tar.gz
+ tar xf tfgrid-sdk-go_Linux_x86_64.tar.gz farmerbot
+ ```
+- Make a copy of the old version in case you need it in the future:
+ ```
+ mv /usr/local/bin/farmerbot /usr/local/bin/farmerbot_archive
+ ```
+- Move the new Farmerbot to the local bin
+ ```
+ mv farmerbot /usr/local/bin
+ ```
+- Restart the bot
+ ```
+ systemctl restart farmerbot
+ ```
+- Remove the tar file
+ ```
+ rm tfgrid-sdk-go_Linux_x86_64.tar.gz
+ ```
+
+# Questions and Feedback
+
+If you have questions concerning the Farmerbot, feel free to ask for help on the [ThreeFold Forum](https://forum.threefold.io/) or on the [ThreeFold Farmer chat](https://t.me/threefoldfarmers).
\ No newline at end of file
diff --git a/collections/documentation/farmers/farmerbot/farmerbot_intro.md b/collections/documentation/farmers/farmerbot/farmerbot_intro.md
new file mode 100644
index 0000000..96d19b3
--- /dev/null
+++ b/collections/documentation/farmers/farmerbot/farmerbot_intro.md
@@ -0,0 +1,15 @@
+ Farmerbot
+
+The Farmerbot is a service that farmers can run in order to automatically manage the nodes in their farms. The behavior of the farmerbot is customizable through a YAML configuration file.
+
+We present here a quick guide to accompany farmers in setting up the Farmerbot. This guide contains the essential information to deploy the Farmerbot on the TFGrid. The other section contains additional information and details on the working of the Farmerbot.
+
+For more information on the Farmerbot, you can visit the [Farmerbot repository](https://github.com/threefoldtech/tfgrid-sdk-go/tree/development/farmerbot) on Github. You can also consult the Farmerbot FAQ if needed.
+
+ Table of Contents
+
+- [Quick Guide](./farmerbot_quick.md)
+- [Additional Information](./farmerbot_information.md)
+- [Minting and the Farmerbot](./farmerbot_minting.md)
+
+> Note: The Farmerbot is an optional feature developed by ThreeFold. Please use at your own risk. While ThreeFold will do its best to fix any issues with the Farmerbot and minting, if minting is affected by the use of the Farmerbot, ThreeFold cannot be held responsible.
\ No newline at end of file
diff --git a/collections/documentation/farmers/farmerbot/farmerbot_minting.md b/collections/documentation/farmers/farmerbot/farmerbot_minting.md
new file mode 100644
index 0000000..f45d13a
--- /dev/null
+++ b/collections/documentation/farmers/farmerbot/farmerbot_minting.md
@@ -0,0 +1,26 @@
+ Minting and the Farmerbot
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Minting Rules](#minting-rules)
+- [Disclaimer](#disclaimer)
+
+***
+
+## Introduction
+
+We cover essential features of ThreeFold minting in relation with the Farmerbot.
+
+## Minting Rules
+
+There are certain minting rules that are very important when it comes to farming on the ThreeFold Grid while using the Farmerbot.
+
+- The 3Node should wake up within 30 minutes of setting the power target to **Up**.
+ - If the 3Node does not respect this rule, the 3Node won't mint for the whole minting period.
+- The 3Node must wake up at least once every 24 hours.
+ - If the 3Node does not respect this rule, the 3Node won't mint for a 24-hour period.
+
+## Disclaimer
+
+Please note that the Farmerbot is an optional feature developed by ThreeFold. Please use at your own risk. While ThreeFold will do its best to fix any issues with the Farmerbot and minting, if minting is affected by the use of the Farmerbot, ThreeFold cannot be held responsible.
diff --git a/collections/documentation/farmers/farmerbot/farmerbot_quick.md b/collections/documentation/farmers/farmerbot/farmerbot_quick.md
new file mode 100644
index 0000000..4177ef8
--- /dev/null
+++ b/collections/documentation/farmers/farmerbot/farmerbot_quick.md
@@ -0,0 +1,292 @@
+ Farmerbot Quick Guide
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Prerequisites](#prerequisites)
+- [Farmerbot Costs on the TFGrid](#farmerbot-costs-on-the-tfgrid)
+- [Enable Wake-On-Lan](#enable-wake-on-lan)
+- [Deploy a Full VM](#deploy-a-full-vm)
+- [Farmerbot Setup](#farmerbot-setup)
+ - [Download the Farmerbot Binaries](#download-the-farmerbot-binaries)
+ - [Create the Farmerbot Files](#create-the-farmerbot-files)
+ - [Run the Farmerbot](#run-the-farmerbot)
+ - [Set a systemd Service](#set-a-systemd-service)
+ - [Check the Farmerbot Logs](#check-the-farmerbot-logs)
+ - [Stop the Farmerbot](#stop-the-farmerbot)
+- [Farmerbot Files](#farmerbot-files)
+ - [Configuration File Template (config.yml)](#configuration-file-template-configyml)
+ - [Environment Variables File Template (.env)](#environment-variables-file-template-env)
+- [Running Multiple Farmerbots on the Same VM](#running-multiple-farmerbots-on-the-same-vm)
+- [Questions and Feedback](#questions-and-feedback)
+
+***
+
+## Introduction
+
+In this guide, we show how to deploy the [Farmerbot](https://github.com/threefoldtech/tfgrid-sdk-go/tree/development/farmerbot) on a full VM running on the TFGrid.
+
+This guide can be done on bare metal or on a full VM running on the TFGrid. You need at least two 3Nodes on the same farm to make use of the Farmerbot.
+
+This version of the Farmerbot also works with ARM64. This means that if you have a Pi 3, 4, or Zero 2 with a 64 bit OS, you can download the appropriate release archive and it will work properly.
+
+Read the [Additional Information](farmerbot_information.md) section for further details concerning the Farmerbot.
+
+## Prerequisites
+
+- The TFChain account associated with the farm should have at least 5 TFT (recommended is 50 TFT)
+
+## Farmerbot Costs on the TFGrid
+
+If you run the Farmerbot on a 3Node on the TFGrid, you will have to pay TFT to deploy on that 3Node. You can run a full VM at minimum specs for the Farmerbot, that is 1vcore, 15GB of SSD storage and 512MB of RAM. Note that you can use the Planetary Network. You do not need to deploy a 3Node with IPv4. The cost on main net for this kind of workload is around 0.175TFT/hour (as of the date 11-07-23).
+
+Next to that, you will have to pay the transaction fees every time the Farmerbot has to wake up or shut down a node. This means that you need some TFT on the account tied to the twin of your farm.
+
+For the periodic wakeups, each node in the farm is shut down and powered on once a day, i.e. 30 times per month. Also, there is 10 random wakeups per month for each node. This means that each node is turned off and on 40 times per month in average. In that case, the average cost per month to power on nodes and shut them back down equals:
+
+> average transaction fees cost per month = 0.001 TFT (extrinsic fee) * amount of nodes * 40 * 2 (1 for powering down, one for powering up)
+
+## Enable Wake-On-Lan
+
+For a 3Node to work properly with the Farmerbot, the parameter wake-on-lan must be enabled. Enabling wake-on-lan on your 3Node may differ depending on your computer model. Please refer to the documentation of your computer if needed.
+
+Usually the feature will be called Wake-on-Lan and you need to set it as "enabled" in the BIOS/UEFI settings.
+
+Here are some examples to guide you:
+
+* Racker Server, Dell R720
+ * Go into `System Setup -> Device Settings -> NIC Port -> NIC Configuration`
+ * Set Wake-on-Lan to `Enable`
+* Desktop Computer, HP EliteDesk G1
+ * Go to Power -> Hardware Power Management
+ * Disable `S5 Maximum Power Saving`
+ * Go to `Advanced -> Power-On Options`
+ * Set `Remote Wake up Boot source` to `Remote Server`
+
+> Hint: Check the Z-OS monitor screen and make sure that all the 3Nodes are within the same lan (e.g. all 3Nodes addresses are between 192.168.15.00 and 192.168.15.255).
+
+For more information on WOL, [read this section](farmerbot_information.md#how-to-prepare-your-farm-for-the-farmerbot-with-wol).
+
+## Deploy a Full VM
+
+For this guide, we run the Farmerbot on a Full VM running on the TFGrid. Note that while you do not need to run the Farmerbot on the TFGrid, the whole process is very simple as presented here.
+
+- Deploy a full VM on the TFGrid
+- Update and upgrade the VM
+ ```
+ apt update && apt upgrade
+ ```
+- Reboot and reconnect to the VM
+ ```
+ reboot
+ ```
+
+## Farmerbot Setup
+
+We present the different steps to run the Farmerbot using the binaries.
+
+> For a script that can help automate the steps in this guide, [check this forum post](https://forum.threefold.io/t/new-farmerbot-install-script/4207).
+
+### Download the Farmerbot Binaries
+
+- Download the latest [ThreeFold tfgrid-sdk-go release](https://github.com/threefoldtech/tfgrid-sdk-go/releases) and extract the farmerbot for your specific setup (here we use `x86_64`). On the line `wget ...`, make sure to replace `` with the latest Farmerbot release.
+ ```
+ wget https://github.com/threefoldtech/tfgrid-sdk-go/releases/download//tfgrid-sdk-go_Linux_x86_64.tar.gz
+ tar xf tfgrid-sdk-go_Linux_x86_64.tar.gz farmerbot
+ ```
+- Move the Farmerbot
+ ```
+ mv farmerbot /usr/local/bin
+ ```
+- Remove the tar file
+ ```
+ rm tfgrid-sdk-go_Linux_x86_64.tar.gz
+ ```
+
+### Create the Farmerbot Files
+
+- Create Farmerbot files directory
+ ```
+ cd ~
+ mkdir farmerbotfiles
+ ```
+- Create the Farmerbot `config.yml` file ([see template below](#configuration-file-template-configyml))
+ ```
+ nano ~/farmerbotfiles/config.yml
+ ```
+- Create the environment variables file and set the variables ([see template below](#environment-variables-file-template-env))
+ ```
+ nano ~/farmerbotfiles/.env
+ ```
+
+### Run the Farmerbot
+
+We run the Farmerbot with the following command:
+
+```
+farmerbot run -e ~/farmerbotfiles/.env -c ~/farmerbotfiles/config.yml -d
+```
+
+For farmers with **ed25519** keys, the flag `-k` should be used. Note that by default, the Farmerbot uses the **sr25519** keys.
+
+```
+farmerbot run -k ed25519 -e ~/farmerbotfiles/.env -c ~/farmerbotfiles/config.yml -d
+```
+
+For more information on the supported commands, the [Additional Information section](farmerbot_information.md#supported-commands-and-flags). You can also consult the [Farmerbot repository](https://github.com/threefoldtech/tfgrid-sdk-go/tree/development/farmerbot).
+
+Once you've verified that the Farmerbot runs properly, you can stop the Farmerbot and go to the next section to set a Farmerbot service. This step will ensure the Farmerbot keeps running after exiting the VM.
+
+### Set a systemd Service
+
+It is highly recommended to set a Ubuntu systemd service to keep the Farmerbot running after exiting the VM.
+
+* Create the service file
+ * ```
+ nano /etc/systemd/system/farmerbot.service
+ ```
+* Set the Farmerbot systemd service
+
+ ```
+ [Unit]
+ Description=ThreeFold Farmerbot
+ StartLimitIntervalSec=0
+
+ [Service]
+ Restart=always
+ RestartSec=5
+ StandardOutput=append:/root/farmerbotfiles/farmerbot.log
+ StandardError=append:/root/farmerbotfiles/farmerbot.log
+ ExecStart=/usr/local/bin/farmerbot run -e /root/farmerbotfiles/.env -c /root/farmerbotfiles/config.yml -d
+
+ [Install]
+ WantedBy=multi-user.target
+ ```
+* Enable the Farmerbot service
+ ```
+ systemctl daemon-reload
+ systemctl enable farmerbot
+ systemctl start farmerbot
+ ```
+* Verify that the Farmerbot service is properly running
+ ```
+ systemctl status farmerbot
+ ```
+
+### Check the Farmerbot Logs
+
+Once you've set a Farmerbot systemd service [as show above](#set-a-systemd-service), the Farmerbot will start writing logs to the file `farmerbot.log` in the directory `farmerbotfiles`.
+
+Thus, you can get more details on the operation of the Farmerbot by inspecting the log file. This can also be used to see the **Farmerbot Report Table** as this table is printed in the Farmerbot log.
+
+* See all logs so far
+ ```
+ cat ~/farmerbotfiles/farmerbot.log
+ ```
+* See the last ten lines and new logs as they are generated
+ ```
+ tail -f ~/farmerbotfiles/farmerbot.log
+ ```
+* See all logs and new lines as they are generated
+ ```
+ tail -f -n +1 ~/farmerbotfiles/farmerbot.log
+ ```
+* See the last report table
+ ```
+ tac ~/farmerbotfiles/farmerbot.log | grep -B5000 -m1 "Nodes report" | tac
+ ```
+
+### Stop the Farmerbot
+
+You can stop the farmerbot with the following command:
+
+```
+systemctl stop farmerbot
+```
+
+After stopping the farmerbot, any nodes in standby mode will remain in standby. To bring them online, use this command:
+
+```
+farmerbot start all -e /root/farmerbotfiles/.env --farm
+```
+
+## Farmerbot Files
+
+### Configuration File Template (config.yml)
+
+In this example, the farm ID is 1, we are setting the Farmerbot with 4 nodes and the node 1 never shuts down, we set a periodic wakeup at 1:00PM.
+
+Note that the timezone of the farmerbot will be the same as the time zone of the machine the farmerbot running inside. By default, a full VM on the TFGrid will be set in UTC.
+
+```
+farm_id: 1
+included_nodes:
+ - 1
+ - 2
+ - 3
+ - 4
+never_shutdown_nodes:
+ - 1
+power:
+ periodic_wake_up_start: 01:00PM
+```
+
+Note that if the user wants to include all the nodes within a farm, they can simply omit the `included_nodes` section. In this case, all nodes of the farm will be included in the Farmerbot, as shown in the example below:
+
+```
+farm_id: 1
+never_shutdown_nodes:
+ - 1
+power:
+ periodic_wake_up_start: 01:00PM
+```
+
+For more information on the configuration file, refer to the [Additional Information section](farmerbot_information.md#yaml-configuration-file-template).
+
+You can also consult the [Farmerbot repository](https://github.com/threefoldtech/tfgrid-sdk-go/tree/development/farmerbot).
+
+### Environment Variables File Template (.env)
+
+The network can be either `main`, `tets`, `dev` or `qa`. The following example is with the main network.
+
+```
+MNEMONIC_OR_SEED="word1 word2 word3 ... word12"
+NETWORK="main"
+```
+
+## Running Multiple Farmerbots on the Same VM
+
+You can run multiple instances of the Farmerbot on the same VM.
+
+To do so, you need to create a directory for each instance of the Farmerbot. Each directory should contain the configuration and variables files as shown above. Once you've set the files, you can simply execute the Farmerbot `run` command to start each bot in each directory.
+
+It's recommended to use distinct names for the directories and the services to easily differentiate the multiple farmerbots running on the VM.
+
+For example, the directory tree of two Farmerbots could be:
+
+```
+└── farmerbotfiles
+ ├── farmerbot1
+ │ ├── .env
+ │ └── config.yml
+ └── farmerbot2
+ ├── .env
+ └── config.yml
+```
+
+For example, the services of two Farmerbots could be named as follows:
+
+```
+farmerbot1.service
+farmerbot2.service
+```
+
+## Questions and Feedback
+
+This guide is meant to get you started quickly with the Farmerbot. That being said, there is a lot more that can be done with the Farmerbot.
+
+For more information on the Farmerbot, please refer to the [Additional Information section](./farmerbot_information.md). You can also consult the [official Farmerbot Go repository](https://github.com/threefoldtech/tfgrid-sdk-go/tree/development/farmerbot).
+
+If you have any questions, you can ask the ThreeFold community for help on the [ThreeFold Forum](https://forum.threefold.io/) or on the [ThreeFold Farmers Chat](https://t.me/threefoldfarmers) on Telegram.
+
+> This is the new version of the Farmerbot written in Go. If you have any feedback and issues, please let us know!
\ No newline at end of file
diff --git a/collections/documentation/farmers/farmerbot/img/farmerbot_4.png b/collections/documentation/farmers/farmerbot/img/farmerbot_4.png
new file mode 100644
index 0000000..58d4b2c
Binary files /dev/null and b/collections/documentation/farmers/farmerbot/img/farmerbot_4.png differ
diff --git a/collections/documentation/farmers/farmerbot/img/farmerbot_5.png b/collections/documentation/farmers/farmerbot/img/farmerbot_5.png
new file mode 100644
index 0000000..439022d
Binary files /dev/null and b/collections/documentation/farmers/farmerbot/img/farmerbot_5.png differ
diff --git a/collections/documentation/farmers/farmerbot/img/farmerbot_bios_1.jpeg b/collections/documentation/farmers/farmerbot/img/farmerbot_bios_1.jpeg
new file mode 100644
index 0000000..7978e9c
Binary files /dev/null and b/collections/documentation/farmers/farmerbot/img/farmerbot_bios_1.jpeg differ
diff --git a/collections/documentation/farmers/farmerbot/img/farmerbot_bios_2.jpeg b/collections/documentation/farmers/farmerbot/img/farmerbot_bios_2.jpeg
new file mode 100644
index 0000000..207aaef
Binary files /dev/null and b/collections/documentation/farmers/farmerbot/img/farmerbot_bios_2.jpeg differ
diff --git a/collections/documentation/farmers/farmerbot/img/farmerbot_bios_3.png b/collections/documentation/farmers/farmerbot/img/farmerbot_bios_3.png
new file mode 100644
index 0000000..10ae6c9
Binary files /dev/null and b/collections/documentation/farmers/farmerbot/img/farmerbot_bios_3.png differ
diff --git a/collections/documentation/farmers/farmers.md b/collections/documentation/farmers/farmers.md
new file mode 100644
index 0000000..9d5f2f8
--- /dev/null
+++ b/collections/documentation/farmers/farmers.md
@@ -0,0 +1,36 @@
+# ThreeFold Farmers
+
+This section covers all practical information on how to become a cloud service provider (farmer) on the ThreeFold Grid.
+
+For complementary information on ThreeFold farming, refer to the [Farming](../../knowledge_base/farming/farming_toc.md) section.
+
+To buy a certified node from an official ThreeFold vendor, check the [ThreeFold Marketplace](https://marketplace.3node.global/).
+
+ Table of Contents
+
+- [Build a 3Node](./3node_building/3node_building.md)
+ - [1. Create a Farm](./3node_building/1_create_farm.md)
+ - [2. Create a Zero-OS Bootstrap Image](./3node_building/2_bootstrap_image.md)
+ - [3. Set the Hardware](./3node_building/3_set_hardware.md)
+ - [4. Wipe All the Disks](./3node_building/4_wipe_all_disks.md)
+ - [5. Set the BIOS/UEFI](./3node_building/5_set_bios_uefi.md)
+ - [6. Boot the 3Node](./3node_building/6_boot_3node.md)
+- [Farming Optimization](./farming_optimization/farming_optimization.md)
+ - [GPU Farming](./3node_building/gpu_farming.md)
+ - [Set Additional Fees](./farming_optimization/set_additional_fees.md)
+ - [Minting Receipts](./3node_building/minting_receipts.md)
+ - [Minting Periods](./farming_optimization/minting_periods.md)
+ - [Room Parameters](./farming_optimization/farm_room_parameters.md)
+ - [Farming Costs](./farming_optimization/farming_costs.md)
+ - [Calculate Your ROI](./farming_optimization/calculate_roi.md)
+- [Advanced Networking](./advanced_networking/advanced_networking_toc.md)
+ - [Networking Overview](./advanced_networking/networking_overview.md)
+ - [Network Considerations](./advanced_networking/network_considerations.md)
+ - [Network Setup](./advanced_networking/network_setup.md)
+- [Farmerbot](./farmerbot/farmerbot_intro.md)
+ - [Quick Guide](./farmerbot/farmerbot_quick.md)
+ - [Additional Information](./farmerbot/farmerbot_information.md)
+ - [Minting and the Farmerbot](./farmerbot/farmerbot_minting.md)
+- [Farmers FAQ](../faq/faq.md#farmers-faq)
+
+> Note: Bugs in the code (e.g. ZOS or other components) can happen. If this is the case, there might be a loss of tokens during minting which won't be refunded by ThreeFold. If there are minting code errors, ThreeFold will try its best to fix the minting code and remint nodes that were affected by such errors.
diff --git a/collections/documentation/farmers/farming_optimization/calculate_roi.md b/collections/documentation/farmers/farming_optimization/calculate_roi.md
new file mode 100644
index 0000000..df13bd4
--- /dev/null
+++ b/collections/documentation/farmers/farming_optimization/calculate_roi.md
@@ -0,0 +1,17 @@
+ Calculate the ROI of a DIY 3Node
+
+To calculate the ROI of a DIY 3Node, we first calculate the Revenue per Month:
+
+>Revenue per month = TFT price when sold * TFT farmed per month
+
+The ROI of a DIY 3Node is:
+
+> Cost of 3Node / Revenue per month = ROI in months
+
+For example, a Rack Server farming 3000 TFT per month with an initial cost of 1500$USD has the following ROI:
+
+> 1500 / (3000 * 0.08) = 6.25 month ROI
+
+This calculation is based on a TFT value of 8 cents. You should adjust this according to the current market price.
+
+Note that this ROI equation is used to compare efficienty between different DIY 3Nodes. It does not constitute real final gains as additional costs must be taken into consideration, such as electricity for the 3Nodes, for the AC system, as well as Internet bandwidth. All those notions are covered in this part of the book.
\ No newline at end of file
diff --git a/collections/documentation/farmers/farming_optimization/farm_room_parameters.md b/collections/documentation/farmers/farming_optimization/farm_room_parameters.md
new file mode 100644
index 0000000..823d84e
--- /dev/null
+++ b/collections/documentation/farmers/farming_optimization/farm_room_parameters.md
@@ -0,0 +1,139 @@
+
+ Air Conditioner, Relative Humidity and Air Changes per Hour
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Calculate the Minimum BTU/h Needed for the AC](#calculate-the-minimum-btuh-needed-for-the-ac)
+ - [How Much BTU/h is Needed?](#how-much-btuh-is-needed)
+ - [Taking Utilization Into Account](#taking-utilization-into-account)
+ - [The General BTU/h Equation](#the-general-btuh-equation)
+- [Ensure Proper Relative Humidity](#ensure-proper-relative-humidity)
+- [Ensure Proper Air Changes per Hour](#ensure-proper-air-changes-per-hour)
+
+***
+
+## Introduction
+
+In this section of the ThreeFold Farmers book, we cover some important notions concerning the room parameters where your 3Nodes are working. We discuss topics such as air conditioner, relative humidity and air changes per hour.
+
+Planning ahead the building of your ThreeFold farm with these notions in mind will ensure a smooth farming experience.
+
+
+
+## Calculate the Minimum BTU/h Needed for the AC
+
+Let's see how to calculate how powerful your AC unit needs to be when it comes to cooling down your server room.
+
+As we know, servers generate heat when they are working. While a desktop 3Node will generate under 20W at idle and a server 3Node might use 100W at **idle**, when you pile up some 3Nodes desktops/servers in the same location, things can get pretty warm when cultivation on the Grid is happening. Indeed, when your servers will be using a lot of power, especially in the summer time, you might need some additional cooling.
+
+A good thing about servers generating heat is that this can be used as a **heat source in the winter**. Other more advanced techniques can be used to maximize the heat production. But that's for another day!
+
+Note that for small farms, your current heating and cooling system may suffice.
+
+So let's do the calculation:
+
+### How Much BTU/h is Needed?
+
+
+How much BTU/h does your ThreeFold Farm need to cool your servers?
+
+Calculating this is pretty simple actually. You need to keep in mind that **1 kW (1000 W) of power is equivalent to 3413 BTU/h** (Britisth Thermal Unit).
+
+> 1000 W = 1 kW = 3413 BTU/h
+>
+> 1000 Wh = 1 kWh = 3413 BTU
+
+So with our idle server example running at 100W, we have 0.1 kW.
+
+> 100 W = 0.1 kW
+
+We then multiply our kW by the BTU/h factor **3413** to obtain the result in BTU/h. Here we have 341.3 BTU/h:
+
+> 0.1 kW * 3413 = 341.3 BTU/h
+
+Say you have 5 servers with this same configuration. It means you have
+
+> (# of servers) * (BTU/h per server) = Total BTU/h
+
+> 5 * 341.3 = 1706.5 BTU/h
+
+Thus, a 2000 BTU/h air conditioner would be able to compensate for the heat when your servers are at idle.
+
+> Note that in general for air conditioners, it will often be written BTU instead of BTU/h as a shorthand.
+
+
+Please take note that this does not take into account the energy needed to cool down your environment. You'd need to take into consideration **the heat of the servers and the general heat of your environment** to figure out how much BTU your AC needs in the big heat days of the summer.
+
+### Taking Utilization Into Account
+
+But then, what happens at cultivation? Well, say your server needs 400W of power when it's being fully cultivated by some lively ThreeFold Users of the New Internet. In this case, we would say that 400 W is the power consumption at **full load**.
+
+As we started with 100 W, and we now have 400 W, it means that you'd need four times the amount of BTU/h.
+
+Here we show how to calculate this with any other configuration of full load/idle.
+
+> Full-load / Idle Ratio = Full Load W / Idle W
+
+> 4 = 400 W / 100 W
+
+The BTU/h needed in cultivation would be
+
+> (Full-Load / Idle Ratio) * Idle BTU/h needed = Full Load BTU/h
+
+> 4 * (1706.5 BTU/h at Idle) = 6826 BTU/h at Full Load
+
+Thus, you would need 6826 BTU/h from the AC unit for 5 servers running each at 400W. In that case, a 8000 BTU/h AC unit would be sufficient. Let's say your environment would typically need 4000 BTU/h to cool the room, you'd need about 12000 BTU/h AC unit for the whole setup.
+
+> If: BTU/h needed < BTU/h AC Unit, Then: AC Unit is OK for TF farming at full load.
+
+
+
+Now you can have a better idea of how much BTU/h is necessary for your AC unit. Of course, this can be a useful piece of data to incorporate in your simulation of Revenue/Cost farming.
+
+### The General BTU/h Equation
+
+The **general equation** would then be:
+
+> Server Power in kW at Full Load * 3413 * Number of Servers = Total Maximum BTU/h needed per ThreeFold Farm
+
+
+
+As another example, 7 servers using 120 W of power at idle would need:
+
+> 0.12 * 3413 * 7 = 2866.92 BTU/h
+
+During cultivation, these 7 servers might use 480 W. This would be:
+
+> 0.48 * 3413 * 7 = 11467.68 BTU/h
+
+To be sure everything's OK, this set up would need a 12 000 BTU/h AC unit to compensate for the heat generated by the ThreeFold Farm during full cultivation. This example considers the environment heat to be negligible.
+
+> 11467.68 < 12000 --> 12K BTU/h AC Unit is OK for farm
+
+
+That's it! It ain't any more complicated. Straight up mathematics and some judgment.
+
+Now, let's compute the costs of running all this!
+
+
+
+## Ensure Proper Relative Humidity
+
+To ensure that the relative humidity in your server room stays within a proper range, look in your server's user manual to know the proper range of relative humidity your server can handle. If necessary, use an hygrometer to measure relative humidity and make sure it stays within an acceptable range for your 3Nodes.
+
+Depending on your geographical location and your current situation, it could be interesting to consider having a AC unit equipped with a dehumidifier. Read your servers' manual to check the proper relative humidity range and set the unit accordingly. The maximum/minimum temperature and relative humidity a 3Node server can handle will depend on the specific server/computer you are using. You should check the server's technical guide/manual to get the proper information. The following is an example.
+
+We will use here the Dell R720 as an example since it is a popular 3Node choice. In this case, we use the R720's [Technical Guide](https://downloads.dell.com/manuals/all-products/esuprt_ser_stor_net/esuprt_poweredge/poweredge-r720_reference-guide_en-us.pdf) as reference.
+
+For the R720, between 35˚C and 40˚C (or 95˚F and 104˚F), with 5% to 85% relative humidity, you can have this <10% of annual operating hours (around 36 days per year), and between 40˚C and 45˚C (or 104˚F and 113˚F), with 5 to 90% relative humidity, it’s <1% of annual operating hours (around 3.6 day per year). All this considers that there is no direct sunlight.
+
+From 10˚C to 35˚C (thus from 50˚F to 95˚F), it’s considered standard operating temperature. With relative humidity from 10% to 80%.
+
+This can give you a good idea of the conditions a 3Node can handle, but make sure you verify with your specific server's manual.
+
+## Ensure Proper Air Changes per Hour
+
+To ensure that the air changes per hour is optimal in your 3Node servers' room, and depending on your current situation, it can be recommended to ventilate the server room in other to disperse or evacuate excess heat and humidity. In those cases, ventilation flow will be set depending on the air changes per hour (ACPH) needed. Note that the [ASHRAE](https://www.ashrae.org/File%20Library/Technical%20Resources/Standards%20and%20Guidelines/Standards%20Addenda/62-2001/62-2001_Addendum-n.pdf) recommends from 10 to 15 ACPH for a computer room.
+
+> Note: A good AC unit will be able to regulate the heat and the relative humidity as well as ensure proper air changes per hour.
\ No newline at end of file
diff --git a/collections/documentation/farmers/farming_optimization/farming_costs.md b/collections/documentation/farmers/farming_optimization/farming_costs.md
new file mode 100644
index 0000000..f10e6a8
--- /dev/null
+++ b/collections/documentation/farmers/farming_optimization/farming_costs.md
@@ -0,0 +1,206 @@
+ Calculate the Farming Costs: Power, Internet and Total Costs
+
+ Table of Contents
+
+- [Calculate the Total Electricity Cost of Your Farm](#calculate-the-total-electricity-cost-of-your-farm)
+- [Calculate the Proper Bandwidth Needed for Your Farm](#calculate-the-proper-bandwidth-needed-for-your-farm)
+ - [The Minimum Bandwidth per 3Node Equation](#the-minimum-bandwidth-per-3node-equation)
+ - [Cost per Month for a Given Bandwidth](#cost-per-month-for-a-given-bandwidth)
+- [Calculate Total Cost and Revenue](#calculate-total-cost-and-revenue)
+ - [Check Revenue with the ThreeFold Simulator](#check-revenue-with-the-threefold-simulator)
+ - [Economics of Farming](#economics-of-farming)
+- [Questions and Feedback](#questions-and-feedback)
+
+***
+
+## Calculate the Total Electricity Cost of Your Farm
+
+The total electricity cost of your farm is the sum of all Power used by your system times the price you pay for each kWh of power.
+
+> Total electricity cost = Total Electricity in kWh * Cost per kWh
+
+> Total Electricty in kWh = 3Nodes' electricity consumption * Number of 3Nodes + Cooling system electricity consumption
+
+With our example, we have 5 servers running at 400 W at Full Load and we have a 12K BTU unit that is consuming in average 1000W.
+
+We would then have:
+
+> 5 * 400 W + 1000 W = 3000 W = 3 kW
+
+To get the kWh per day we simply multiply by 24.
+
+> kW * (# of hour per day) = daily kWh consumption
+
+> 3 kW * 24 = 72 kWh / day
+
+We thus have 72 kWH per day. For 30 days, this would be
+
+> kWh / day * (# day in a month) = kWh per month
+
+> 72 * 30 = 2160 kWH / month.
+
+At a kWh price of 0.10$ USD, we have a cost of 216 $USD per month for the electricity bill of our ThreeFold farm.
+
+> kWH / month of the farm * kWh Cost = Electricity Bill per month for the farm
+
+> 2160 * 0.1 = 216$USD / month for electricity bills
+
+
+## Calculate the Proper Bandwidth Needed for Your Farm
+
+The bandwidth needed for a given 3Node is not yet set in stone and you are welcome to participate in ongoing [discussion on this subject](https://forum.threefold.io/t/storage-bandwidth-ratio/1389) on the ThreeFold Forum.
+
+In this section, we will give general guidelines. The goal is to have a good idea of what constitutes a proper bandwidth available for a given amount of resources utilized on the ThreeFold Grid.
+
+Starting with a minimum of 1 mbps per Titan, which is 1 TB SSD and 32 GB RAM, we note that this is the lowest limit that gives the opportunity for the most people possible to join the ThreeFold Grid. That being said, we could set that 10 mbps is an acceptable upper limit for 1 TB SSD and 64 GB of RAM.
+
+Those numbers are empirical and more information will be shared in the future. The ratio 1TB SSD/64GB RAM is in tune with the optimal TFT rewards ratio. It is thus logical to think that farmers will build 3Node based on this ratio. Giving general bandwidth guidelines based on this ratio unit could thus be efficient for the current try-and-learn situation.
+
+### The Minimum Bandwidth per 3Node Equation
+
+
+Here we explore some equations that can give a general idea to farmers of the bandwidth needed for their farms. As stated, this is not yet set in stones and the TFDAO will need to discuss and clarify those notions.
+
+Here is a general equation that gives you a good idea of a correct bandwidth for a 3Node:
+
+> min Bandwidth per 3Node (mbps) = k * max((Total SSD TB / 1 Tb),(Total Threads / 8 Threads),(Total GB / 64 GB)) + k * (Total HDD TB / 2)
+
+Setting k = 10 mbps, we have:
+
+> min Bandwidth per 3Node (mbps) = 10 * max((Total SSD TB / 1 TB),(Total Threads / 8 Threads),(Total GB / 64 GB)) + 10 * (Total HDD TB / 2)
+
+As an example, a Titan, with 1TB SSD, 8 Threads and 64 GB of RAM, would need 10 mbps:
+
+> 10 * max(1, 1, 1) = 10 * 1 = 10
+
+With the last portion of the equation, we can see that for each additional 1TB HDD storage, you would need to add 5 mbps of bandwidth.
+
+
+Let's take a big server as another example. Say we have a server with 5TB SSD, 48 threads and 384 GB of RAM. We would then need 60 mbps of bandwidth for each of these 3Nodes:
+
+> 10 * max((5/5), (48/8), (384/64)) = 10 * max(5,6,6) = 10 * 6 = 60
+
+This server would need 60 mbps minimum to account for a full TF Grid utilization.
+
+You can easily scale this equation if you have many 3Nodes.
+
+
+
+Let's say you have a 1 gbps bandwidth from your Internet Service Provider (ISP). How much of those 3Nodes could your farm have?
+
+> Floor (Total available bandwidth / ((Bandwidth needed per 3Nodes)) = Max servers possible
+
+With our example we have:
+
+> 1000 / 60 = 16.66... = 16
+
+We note that the function Floor takes the integer without the decimals.
+
+Thus, a 1 gbps bandwidth farm could have 16 3Nodes with each 5TB SSD, 48 threads and 384 GB of RAM.
+
+
+
+In this section, we used **k = 10 mbps**. If you follow those guidelines, you will most probably have a decent bandwidth for your ThreeFold farm. For the time being, the goal is to have farmers building ThreeFold farms and scale them reasonably with their available bandwidth.
+
+Stay tuned for official bandwidth parameters in the future.
+
+
+
+### Cost per Month for a Given Bandwidth
+
+Once you know the general bandwidth needed for your farm, you can check with your ISP the price per month and take this into account when calculating your monthly costs.
+
+Let's take the example we used with 5 servers with 400 W at Full Load. Let's say these 5 servers have the same parameters we used above here. We then need 60 gbps per 3Nodes. This means we need 300 mbps. For the sake of our example, let's say this is around 100$ USD per month.
+
+
+## Calculate Total Cost and Revenue
+
+
+As the TFT price is fixed for 60 months when you connect your 3Node for the first time on the TF Grid, we will use the period of 60 months, or 5 years, to calculate the total cost and revenue.
+
+The total cost is equal to:
+
+> Total Cost = Initial investment + 60 * (electricity + Internet costs per month)
+
+In our example, we can state that we paid each server 1500$ USD and that they generate each 3000 TFT per month, with an entry price of 0.08$ USD per TFT.
+
+The electricity cost per month is
+
+> 144$ for the electricity bill
+>
+> 100$ for the Internet bill
+>
+> Total : 244 $ monthly cost for electricity and Internet
+
+The revenues are
+
+> Revenues per month = Number of 3Nodes * TFT farmed per 3Node * Price TFT Sold
+
+In this example, we have 5 servers generating 2000 TFT per month at 0.08$ USD per TFT:
+
+> 5 * 3000$ * 0.08$ = 1200$
+
+The net revenue per month are thus equal to
+
+> Net Revenue = Gross revenue - Monthly cost.
+
+We thus have
+
+> 1200$ - 244$ = 956$
+
+This means that we generate a net profit of 956$ per month, without considering the initial investment of building the 3Nodes for the farm.
+
+In the previous AC example, we calculate that a minimum of 12K BTU was needed for the AC system. Let's say that this would mean buying a 350$ USD 12k BTU AC unit.
+
+The initial cost is the cost of all the 3Nodes plus the AC system.
+
+> Number of 3Nodes * cost per 3Nodes + Cost of AC system = Total Cost
+
+In this case, it would be:
+
+> Total initial investment = Number of 3Nodes * Cost of 3Node + Cost of AC system
+
+Then we'd have:
+
+> 5 * 1500 + 350 = 7850 $
+
+Thus, a more realistic ROI would be:
+
+> Total initial investment / Net Revenue per Month = ROI in months
+
+In our case, we would have:
+
+> 7850$ / 956$ = Ceiling(8.211...) = 9
+
+With the function Ceiling taking the upper integer, without any decimals.
+
+Then within 9 months, this farm would have paid itself and from now on, it would be only positive net revenue of 956$ per month.
+
+We note that this takes into consideration that we are using the AC system 24/7. This would surely not be the case in real life. This means that the real ROI would be even better. It is a common practice to do estimates with stricter parameters. If you predict being profitable with strict parameters, you will surely be profitable in real life, even when "things" happen and not everything goes as planned. As always, this is not financial advice.
+
+We recall that in the section [Calculate the ROI of a DIY 3Node](./calculate_roi.md), we found a simpler ROI of 6.25 months, say 7 months, that wasn't taking into consideration the additional costs of Internet and electricity. We now have a more realistic ROI of 9 months based on a fixed TFT price of 0.08$ USD. You will need to use to equations and check with your current TF farm and 3Nodes, as well as the current TFT market price.
+
+
+### Check Revenue with the ThreeFold Simulator
+
+To know how much TFT you will farm per month for a giving 3Node, the easiest route is to use the [ThreeFold Simulator](https://simulator.grid.tf/). You can do predictions of 60 months as the TFT price is locked at the TFT price when you first connect your 3Node, and this, for 60 months.
+
+To know the details of the calculations behind this simulator, you can read [this documentation](https://library.threefold.me/info/threefold#/tfgrid/farming/threefold__farming_reward).
+
+
+### Economics of Farming
+
+As a brief synthesis, the following equations are used to calculate the total revenues and costs of your farm.
+
+```
+- Total Monthly Cost = Electricity cost + Internet Cost
+- Total Electricity Used = Electricy per 3Node * Number of 3Node + Electricity for Cooling
+- Total Monthly Revenue = TFT farmed per 3 node * Number of 3Nodes * TFT price when sold
+- Initial Investment = Price of farm (3Nodes) + Price of AC system
+- Total Return on investment = (60 * Monthly Revenue) - (60 * Monthly cost) - Initial Investment
+```
+
+
+## Questions and Feedback
+
+This section constitutes a quick synthesis of the costs and revenues when running a ThreeFold Farm. As always, do your own reseaerch and don't hesitate to visit the [ThreeFold Forum](https://forum.threefold.io/) on the [ThreeFold Telegram Farmer Group](https://t.me/threefoldfarmers) if you have any questions.
diff --git a/collections/documentation/farmers/farming_optimization/farming_optimization.md b/collections/documentation/farmers/farming_optimization/farming_optimization.md
new file mode 100644
index 0000000..10f14b0
--- /dev/null
+++ b/collections/documentation/farmers/farming_optimization/farming_optimization.md
@@ -0,0 +1,13 @@
+ Farming Optimization
+
+The section [Build a 3Node](../3node_building/3node_building.md) covered the notions necessary to build a DIY 3Node server. The following section will give you additional information with the goal of optimizing your farm while also being able to plan ahead the costs in terms of energy and capitals. We also cover how to set a GPU node and more.
+
+ Table of Contents
+
+- [GPU Farming](../3node_building/gpu_farming.md)
+- [Set Additional Fees](./set_additional_fees.md)
+- [Minting Receipts](../3node_building/minting_receipts.md)
+- [Minting Periods](./minting_periods.md)
+- [Room Parameters](./farm_room_parameters.md)
+- [Farming Costs](./farming_costs.md)
+- [Calculate Your ROI](./calculate_roi.md)
\ No newline at end of file
diff --git a/collections/documentation/farmers/farming_optimization/minting_periods.md b/collections/documentation/farmers/farming_optimization/minting_periods.md
new file mode 100644
index 0000000..5979f84
--- /dev/null
+++ b/collections/documentation/farmers/farming_optimization/minting_periods.md
@@ -0,0 +1,56 @@
+Minting Periods
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Minting Period Length](#minting-period-length)
+- [2023 Minting Periods](#2023-minting-periods)
+- [2024 Minting Periods](#2024-minting-periods)
+
+***
+
+## Introduction
+
+We discuss the length and the frequencies of the ThreeFold farming minting periods.
+
+## Minting Period Length
+
+Each minting period has: 2630880 seconds = 43848 minutes = 730.8 hours.
+
+## 2023 Minting Periods
+
+The minting periods for the 12 months of 2023 are the following:
+
+| Month | Start of the Minting Period | End of the Minting Period |
+|----------|---------------------------------|---------------------------------|
+| Jan 2023 | December 31, 2022 at 4\:32\:40 am | January 30, 2023 at 3\:20\:40 pm |
+| Feb 2023 | January 30, 2023 at 3\:20\:40 pm | March 2, 2023 at 2\:08\:40 am |
+| Mar 2023 | March 2, 2023 at 2\:08\:40 am | April 1, 2023 at 12\:56\:40 pm |
+| Apr 2023 | April 1, 2023 at 12\:56\:40 pm | May 1, 2023 at 11\:44\:40 pm |
+| May 2023 | May 1, 2023 at 11\:44\:40 pm | June 1, 2023 at 10\:32\:40 am |
+| Jun 2023 | June 1, 2023 at 10\:32\:40 am | July 1, 2023 at 9\:20\:40 pm |
+| Jul 2023 | July 1, 2023 at 9\:20\:40 pm | August 1, 2023 at 8\:08\:40 am |
+| Aug 2023 | August 1, 2023 at 8\:08\:40 am | August 31, 2023 at 6\:56\:40 pm |
+| Sep 2023 | August 31, 2023 at 6\:56\:40 pm | October 1, 2023 at 5\:44\:40 am |
+| Oct 2023 | October 1, 2023 at 5\:44\:40 am | October 31, 2023 at 4\:32\:40 pm |
+| Nov 2023 | October 31, 2023 at 4\:32\:40 pm | December 1, 2023 at 3\:20\:40 am |
+| Dec 2023 | December 1, 2023 at 3\:20\:40 am | December 31, 2023 at 2\:08\:40 pm |
+
+## 2024 Minting Periods
+
+The minting periods for the 12 months of 2024 are the following:
+
+| Month | Start of the Minting Period | End of the Minting Period |
+|----------|---------------------------------|---------------------------------|
+| Jan 2024 | December 31, 2023 at 14\:08\:40 | January 31, 2024 at 00\:56\:40 |
+| Feb 2024 | January 31, 2024 at 00\:56\:40 | March 1, 2024 at 11\:44\:40 |
+| Mar 2024 | March 1, 2024 at 11\:44\:40 | March 31, 2024 at 22\:32\:40 |
+| Apr 2024 | Marc 31, 2024 at 22\:32\:40 | May 1, 2024 at 09\:20\:40 |
+| May 2024 | May 1, 2024 at 09\:20\:40 | May 31, 2024 at 20\:08\:40 |
+| Jun 2024 | May 31, 2024 at 20\:08\:40 | July 1, 2024 at 06\:56\:40 |
+| Jul 2024 | July 1, 2024 at 06\:56\:40 | July 31, 2024 at 17:44\:40 |
+| Aug 2024 | July 31, 2024 at 17\:44\:40 | August 31, 2024 at 04\:32\:40 |
+| Sep 2024 | August 31, 2024 at 04\:32\:40 | September 30, 2024 at 15\:20\:40 |
+| Oct 2024 | September 30, 2024 at 15\:20\:40 | October 31, 2024 at 02\:08\:40 |
+| Nov 2024 | October 31, 2024 at 02\:08\:40 | November 30, 2024 at 12\:56\:40 |
+| Dec 2024 | November 30, 2024 at 12\:56\:40 | December 30, 2024 at 23\:44\:40 |
diff --git a/collections/documentation/farmers/farming_optimization/set_additional_fees.md b/collections/documentation/farmers/farming_optimization/set_additional_fees.md
new file mode 100644
index 0000000..e1b5437
--- /dev/null
+++ b/collections/documentation/farmers/farming_optimization/set_additional_fees.md
@@ -0,0 +1,31 @@
+Set Additional Fees
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Steps](#steps)
+- [TFT Payments](#tft-payments)
+- [Dedicated Nodes Notice](#dedicated-nodes-notice)
+
+***
+
+## Introduction
+
+Farmers can set additional fees for their 3Nodes on the [TF Dashboard](https://dashboard.grid.tf/). By doing so, users will then be able to [reserve the 3Node and use it as a dedicated node](../../dashboard/deploy/dedicated_machines.md).
+This can be useful for farmers who provide additional values to their 3Nodes, e.g. a GPU card and/or high-quality hardware.
+
+## Steps
+
+Here are the steps to [set additional fees](../../dashboard/farms/your_farms.md#extra-fees) to a 3Node.
+
+* On the Dashboard, go to **Farms** -> **Your Farms**
+* Under the section **Your Nodes**, locate the 3Node and click **Set Additional Fees** under **Actions**
+* Set a monthly fee (in USD) and click **Set**
+
+## TFT Payments
+
+When a user reserves your 3Node, you will receive TFT payments once every 24 hours. These TFT payments will be sent to the TFChain account of your farm's twin.
+
+## Dedicated Nodes Notice
+
+Note that while any 3Node that has no workload can be reserved by a TF user as a dedicated node, when a farmer sets additional fees to a 3Node, this 3Node automatically becomes a dedicated node. For a user to run workloads on this 3Node, the 3Node must then be reserved, i.e rented as a dedicated node.
\ No newline at end of file
diff --git a/collections/documentation/farmers/img/farming_30.png b/collections/documentation/farmers/img/farming_30.png
new file mode 100644
index 0000000..d810d13
Binary files /dev/null and b/collections/documentation/farmers/img/farming_30.png differ
diff --git a/collections/documentation/system_administrators/advanced/advanced.md b/collections/documentation/system_administrators/advanced/advanced.md
new file mode 100644
index 0000000..92a95c3
--- /dev/null
+++ b/collections/documentation/system_administrators/advanced/advanced.md
@@ -0,0 +1,14 @@
+ TFGrid Advanced
+
+In this section, we delve into sophisticated topics and powerful functionalities that empower you to harness the full potential of TFGrid 3.0. Whether you're an experienced user seeking to deepen your understanding or a trailblazer venturing into uncharted territories, this manual is your gateway to mastering advanced concepts on the ThreeFold Grid.
+
+Table of Contents
+
+- [Token Transfer Keygenerator](./token_transfer_keygenerator.md)
+- [Cancel Contracts](./cancel_contracts.md)
+- [Contract Bills Reports](./contract_bill_report.md)
+- [Listing Free Public IPs](./list_public_ips.md)
+- [Redis](./grid3_redis.md)
+- [IPFS](./ipfs/ipfs_toc.md)
+ - [IPFS on a Full VM](./ipfs/ipfs_fullvm.md)
+ - [IPFS on a Micro VM](./ipfs/ipfs_microvm.md)
diff --git a/collections/documentation/system_administrators/advanced/cancel_contracts.md b/collections/documentation/system_administrators/advanced/cancel_contracts.md
new file mode 100644
index 0000000..7b466a0
--- /dev/null
+++ b/collections/documentation/system_administrators/advanced/cancel_contracts.md
@@ -0,0 +1,48 @@
+ Cancel Contracts
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Using the Dashboard](#using-the-dashboard)
+- [Using GraphQL and Polkadot UI](#using-graphql-and-polkadot-ui)
+- [Using grid3\_client\_ts](#using-grid3_client_ts)
+
+***
+
+## Introduction
+
+We present different methods to delete contracts on the TFGrid.
+
+## Using the Dashboard
+
+To cancel contracts with the Dashboard, consult the [Contracts List](../../dashboard/deploy/your_contracts.md) documentation.
+
+## Using GraphQL and Polkadot UI
+
+From the QraphQL service execute the following query.
+
+```
+query MyQuery {
+
+ nodeContracts(where: {twinId_eq: TWIN_ID, state_eq: Created}) {
+ contractId
+ }
+}
+
+```
+
+replace `TWIN_ID` with your twin id. The information should be available on the [Dashboard](../../dashboard/dashboard.md).
+
+Then from [polkadot UI](https://polkadot.js.org/apps/), add the tfchain endpoint to development.
+
+![](img/polka_web_add_development_url.png)
+
+Go to `Extrinsics`, choose the `smartContract` module and `cancelContract` extrinsic and use the IDs from GraphQL to execute the cancelation.
+
+![](img/polka_web_cancel_contracts.jpg)
+
+## Using grid3_client_ts
+
+In order to use the `grid3_client_ts` module, it is essential to first clone our official mono-repo containing the module and then navigate to it. If you are looking for a quick and efficient way to cancel contracts, we offer a code-based solution that can be found [here](https://github.com/threefoldtech/tfgrid-sdk-ts/blob/development/packages/grid_client/scripts/delete_all_contracts.ts).
+
+To make the most of `grid_client`, we highly recommend following our [Grid-Client guide](https://github.com/threefoldtech/tfgrid-sdk-ts/blob/development/packages/grid_client/README.md) for a comprehensive overview of the many advanced capabilities offered by this powerful tool. With features like contract creation, modification, and retrieval, `grid_client` provides an intuitive and easy-to-use solution for managing your contracts effectively.
diff --git a/collections/documentation/system_administrators/advanced/contract_bill_report.md b/collections/documentation/system_administrators/advanced/contract_bill_report.md
new file mode 100644
index 0000000..1269df2
--- /dev/null
+++ b/collections/documentation/system_administrators/advanced/contract_bill_report.md
@@ -0,0 +1,63 @@
+ Contract Bills Reports
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Contract Billing Report (GraphQL)](#contract-billing-report-graphql)
+- [Consumption](#consumption)
+
+***
+
+## Introduction
+
+Now you can check the billing rate of your contracts directly from the `Contracts` tab in the Dashboard.
+
+> It takes an hour for the contract to display the billing rate (Until it reaches the first billing cycle).
+
+The `Billing Rate` is displayed in `TFT/Hour`
+
+![image](img/billing_rate.png)
+
+## Contract Billing Report (GraphQL)
+
+- you need to find the contract ID
+- ask the graphql for the consumption
+
+> example query for all contracts
+
+```graphql
+query MyQuery {
+ contractBillReports {
+ contractId
+ amountBilled
+ discountReceived
+ }
+}
+```
+
+And for a specific contract
+
+```graphql
+query MyQuery {
+ contractBillReports(where: { contractId_eq: 10 }) {
+ amountBilled
+ discountReceived
+ contractId
+ }
+}
+```
+
+## Consumption
+
+```graphql
+query MyQuery {
+ consumptions(where: { contractId_eq: 10 }) {
+ contractId
+ cru
+ sru
+ mru
+ hru
+ nru
+ }
+}
+```
diff --git a/collections/documentation/system_administrators/advanced/grid3_redis.md b/collections/documentation/system_administrators/advanced/grid3_redis.md
new file mode 100644
index 0000000..0d3377e
--- /dev/null
+++ b/collections/documentation/system_administrators/advanced/grid3_redis.md
@@ -0,0 +1,46 @@
+ Redis
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Install Redis](#install-redis)
+ - [Linux](#linux)
+ - [MacOS](#macos)
+- [Run Redis](#run-redis)
+
+***
+
+## Introduction
+
+Redis is an open-source, in-memory data structure store that is widely used as a caching layer, message broker, and database. It is known for its speed, versatility, and support for a wide range of data structures. Redis is designed to deliver high-performance data access by storing data in memory, which allows for fast read and write operations. It supports various data types, including strings, lists, sets, hashes, and more, and provides a rich set of commands for manipulating and querying the data.
+
+Redis is widely used in various use cases, including caching, session management, real-time analytics, leaderboards, task queues, and more. Its simplicity, speed, and flexibility make it a popular choice for developers who need a fast and reliable data store for their applications. In Threefold's ecosystem context, Redis can be used as a backend mechanism to communicate with the nodes on the ThreeFold Grid using the Reliable Message Bus.
+
+
+
+## Install Redis
+
+### Linux
+
+If you don't find Redis in your Linux distro's package manager, check the [Redis downloads](https://redis.io/download) page for the source code and installation instructions.
+
+### MacOS
+
+On MacOS, [Homebrew](https://brew.sh/) can be used to install Redis. The steps are as follow:
+
+```
+brew update
+brew install redis
+```
+
+Alternatively, it can be built from source, using the same [download page](https://redis.io/download/) as shown above.
+
+
+
+## Run Redis
+
+You can launch the Redis server with the following command:
+
+```
+redis-server
+```
\ No newline at end of file
diff --git a/collections/documentation/system_administrators/advanced/grid3_stellar_tfchain_bridge.md b/collections/documentation/system_administrators/advanced/grid3_stellar_tfchain_bridge.md
new file mode 100644
index 0000000..f74cbdb
--- /dev/null
+++ b/collections/documentation/system_administrators/advanced/grid3_stellar_tfchain_bridge.md
@@ -0,0 +1,39 @@
+ Transferring TFT Between Stellar and TFChain
+
+Table of Contents
+
+- [Usage](#usage)
+- [Prerequisites](#prerequisites)
+- [Stellar to TFChain](#stellar-to-tfchain)
+- [TFChain to Stellar](#tfchain-to-stellar)
+
+***
+
+## Usage
+
+This document will explain how you can transfer TFT from Tfchain to Stellar and back.
+
+For more information on TFT bridges, read [this documentation](../threefold_token/tft_bridges/tft_bridges.md).
+
+## Prerequisites
+
+- [Stellar wallet](../threefold_token/storing_tft/storing_tft.md)
+
+- [Account on TFChain (use TF Dashboard to create one)](../dashboard/wallet_connector.md)
+
+![](./img/bridge.png)
+
+## Stellar to TFChain
+
+You can deposit to Tfchain using the bridge page on the TF Dashboard, click deposit:
+
+![bridge](./img/bridge_deposit.png)
+
+## TFChain to Stellar
+
+You can bridge back to stellar using the bridge page on the dashboard, click withdraw:
+
+![withdraw](./img/bridge_withdraw.png)
+
+A withdrawfee of 1 TFT will be taken, so make sure you send a larger amount as 1 TFT.
+The amount withdrawn from TFChain will be sent to your Stellar wallet.
diff --git a/collections/documentation/system_administrators/advanced/img/advanced_.png b/collections/documentation/system_administrators/advanced/img/advanced_.png
new file mode 100644
index 0000000..0065213
Binary files /dev/null and b/collections/documentation/system_administrators/advanced/img/advanced_.png differ
diff --git a/collections/documentation/system_administrators/advanced/img/billing_rate.png b/collections/documentation/system_administrators/advanced/img/billing_rate.png
new file mode 100644
index 0000000..b22a35a
Binary files /dev/null and b/collections/documentation/system_administrators/advanced/img/billing_rate.png differ
diff --git a/collections/documentation/system_administrators/advanced/img/bridge.png b/collections/documentation/system_administrators/advanced/img/bridge.png
new file mode 100644
index 0000000..269dead
Binary files /dev/null and b/collections/documentation/system_administrators/advanced/img/bridge.png differ
diff --git a/collections/documentation/system_administrators/advanced/img/bridge_deposit.png b/collections/documentation/system_administrators/advanced/img/bridge_deposit.png
new file mode 100644
index 0000000..97619cc
Binary files /dev/null and b/collections/documentation/system_administrators/advanced/img/bridge_deposit.png differ
diff --git a/collections/documentation/system_administrators/advanced/img/bridge_withdraw.png b/collections/documentation/system_administrators/advanced/img/bridge_withdraw.png
new file mode 100644
index 0000000..461a80d
Binary files /dev/null and b/collections/documentation/system_administrators/advanced/img/bridge_withdraw.png differ
diff --git a/collections/documentation/system_administrators/advanced/img/contracts_list.png b/collections/documentation/system_administrators/advanced/img/contracts_list.png
new file mode 100644
index 0000000..a252229
Binary files /dev/null and b/collections/documentation/system_administrators/advanced/img/contracts_list.png differ
diff --git a/collections/documentation/system_administrators/advanced/img/ipfs_logo.png b/collections/documentation/system_administrators/advanced/img/ipfs_logo.png
new file mode 100644
index 0000000..03a5db4
Binary files /dev/null and b/collections/documentation/system_administrators/advanced/img/ipfs_logo.png differ
diff --git a/collections/documentation/system_administrators/advanced/img/polka_web_add_development_url.png b/collections/documentation/system_administrators/advanced/img/polka_web_add_development_url.png
new file mode 100644
index 0000000..159e7a1
Binary files /dev/null and b/collections/documentation/system_administrators/advanced/img/polka_web_add_development_url.png differ
diff --git a/collections/documentation/system_administrators/advanced/img/polka_web_cancel_contracts.jpg b/collections/documentation/system_administrators/advanced/img/polka_web_cancel_contracts.jpg
new file mode 100644
index 0000000..b704900
Binary files /dev/null and b/collections/documentation/system_administrators/advanced/img/polka_web_cancel_contracts.jpg differ
diff --git a/collections/documentation/system_administrators/advanced/img/swap_to_stellar.png b/collections/documentation/system_administrators/advanced/img/swap_to_stellar.png
new file mode 100644
index 0000000..0473664
Binary files /dev/null and b/collections/documentation/system_administrators/advanced/img/swap_to_stellar.png differ
diff --git a/collections/documentation/system_administrators/advanced/ipfs/ipfs_fullvm.md b/collections/documentation/system_administrators/advanced/ipfs/ipfs_fullvm.md
new file mode 100644
index 0000000..e61a173
--- /dev/null
+++ b/collections/documentation/system_administrators/advanced/ipfs/ipfs_fullvm.md
@@ -0,0 +1,190 @@
+ IPFS on a Full VM
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Deploy a Full VM](#deploy-a-full-vm)
+- [Create a Root-Access User](#create-a-root-access-user)
+- [Set a Firewall](#set-a-firewall)
+ - [Additional Ports](#additional-ports)
+- [Install IPFS](#install-ipfs)
+- [Set IPFS](#set-ipfs)
+- [Final Verification](#final-verification)
+- [Questions and Feedback](#questions-and-feedback)
+
+***
+
+## Introduction
+
+In this ThreeFold guide, we explore how to set an IPFS node on a Full VM using the ThreeFold Playground.
+
+## Deploy a Full VM
+
+We start by deploying a full VM on the ThreeFold Playground.
+
+* Go to the [Threefold Playground](https://playground.grid.tf/#/)
+* Deploy a full VM (Ubuntu 20.04) with an IPv4 address and at least the minimum specs
+ * IPv4 Address
+ * Minimum vcores: 1vcore
+ * Minimum MB of RAM: 1024GB
+ * Minimum storage: 50GB
+* After deployment, note the VM IPv4 address
+* Connect to the VM via SSH
+ * ```
+ ssh root@VM_IPv4_address
+ ```
+
+## Create a Root-Access User
+
+We create a root-access user. Note that this step is optional.
+
+* Once connected, create a new user with root access (for this guide we use "newuser")
+ * ```
+ adduser newuser
+ ```
+ * You should now see the new user directory
+ * ```
+ ls /home
+ ```
+ * Give sudo capacity to the new user
+ * ```
+ usermod -aG sudo newuser
+ ```
+ * Switch to the new user
+ * ```
+ su - newuser
+ ```
+ * Create a directory to store the public key
+ * ```
+ mkdir ~/.ssh
+ ```
+ * Give read, write and execute permissions for the directory to the new user
+ * ```
+ chmod 700 ~/.ssh
+ ```
+ * Add the SSH public key in the file **authorized_keys** and save it
+ * ```
+ nano ~/.ssh/authorized_keys
+ ```
+* Exit the VM
+ * ```
+ exit
+ ```
+* Reconnect with the new user
+ * ```
+ ssh newuser@VM_IPv4_address
+ ```
+
+## Set a Firewall
+
+We set a firewall to monitor and control incoming and outgoing network traffic. To do so, we will define predetermined security rules. As a firewall, we will be using [Uncomplicated Firewall](https://wiki.ubuntu.com/UncomplicatedFirewall) (ufw).
+For our security rules, we want to allow SSH, HTTP and HTTPS (443 and 8443).
+We thus add the following rules:
+* Allow SSH (port 22)
+ * ```
+ sudo ufw allow ssh
+ ```
+* Allow port 4001
+ * ```
+ sudo ufw allow 4001
+ ```
+* To enable the firewall, write the following:
+ * ```
+ sudo ufw enable
+ ```
+* To see the current security rules, write the following:
+ * ```
+ sudo ufw status verbose
+ ```
+You now have enabled the firewall with proper security rules for your IPFS deployment.
+
+### Additional Ports
+
+We provided the basic firewall ports for your IPFS instance. There are other more advanced configurations possible.
+If you want to access your IPFS node remotely, you can allow **port 5001**. This will allow anyone to access your IPFS node. Make sure that you know what you are doing if you go this route. You should, for example, restrict which external IP address can access port 5001.
+If you want to run your deployment as a gateway node, you should allow **port 8080**. Read the IPFS documentation for more information on this.
+If you want to run pubsub capabilities, you need to allow **port 8081**. For more information, read the [IPFS documentation](https://blog.ipfs.tech/25-pubsub/).
+
+## Install IPFS
+
+We install the [IPFS Kubo binary](https://docs.ipfs.tech/install/command-line/#install-official-binary-distributions).
+* Download the binary
+ * ```
+ wget https://dist.ipfs.tech/kubo/v0.24.0/kubo_v0.24.0_linux-amd64.tar.gz
+ ```
+* Unzip the file
+ * ```
+ tar -xvzf kubo_v0.24.0_linux-amd64.tar.gz
+ ```
+* Change directory
+ * ```
+ cd kubo
+ ```
+* Run the install script
+ * ```
+ sudo bash install.sh
+ ```
+* Verify that IPFS Kubo is properly installed
+ * ```
+ ipfs --version
+ ```
+
+## Set IPFS
+
+We initialize IPFS and run the IPFS daemon.
+
+* Initialize IPFS
+ * ```
+ ipfs init --profile server
+ ```
+* Increase the storage capacity (optional)
+ * ```
+ ipfs config Datastore.StorageMax 30GB
+ ```
+* Run the IPFS daemon
+ * ```
+ ipfs daemon
+ ```
+* Set an Ubuntu systemd service to keep the IPFS daemon running after exiting the VM
+ * ```
+ sudo nano /etc/systemd/system/ipfs.service
+ ```
+* Enter the systemd info
+ * ```
+ [Unit]
+ Description=IPFS Daemon
+ [Service]
+ Type=simple
+ ExecStart=/usr/local/bin/ipfs daemon --enable-gc
+ Group=newuser
+ Restart=always
+ Environment="IPFS_PATH=/home/newuser/.ipfs"
+ [Install]
+ WantedBy=multi-user.target
+ ```
+* Enable the service
+ * ```
+ sudo systemctl daemon-reload
+ sudo systemctl enable ipfs
+ sudo systemctl start ipfs
+ ```
+* Verify that the IPFS daemon is properly running
+ * ```
+ sudo systemctl status ipfs
+ ```
+## Final Verification
+We reboot and reconnect to the VM and verify that IPFS is properly running as a final verification.
+* Reboot the VM
+ * ```
+ sudo reboot
+ ```
+* Reconnect to the VM
+ * ```
+ ssh newuser@VM_IPv4_address
+ ```
+* Check that the IPFS daemon is running
+ * ```
+ ipfs swarm peers
+ ```
+## Questions and Feedback
+If you have any questions or feedback, please let us know by either writing a post on the [ThreeFold Forum](https://forum.threefold.io/), or by chatting with us on the [TF Grid Tester Community](https://t.me/threefoldtesting) Telegram channel.
\ No newline at end of file
diff --git a/collections/documentation/system_administrators/advanced/ipfs/ipfs_microvm.md b/collections/documentation/system_administrators/advanced/ipfs/ipfs_microvm.md
new file mode 100644
index 0000000..2f58f16
--- /dev/null
+++ b/collections/documentation/system_administrators/advanced/ipfs/ipfs_microvm.md
@@ -0,0 +1,167 @@
+ IPFS on a Micro VM
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Deploy a Micro VM](#deploy-a-micro-vm)
+- [Install the Prerequisites](#install-the-prerequisites)
+- [Set a Firewall](#set-a-firewall)
+ - [Additional Ports](#additional-ports)
+- [Install IPFS](#install-ipfs)
+- [Set IPFS](#set-ipfs)
+- [Set IPFS with zinit](#set-ipfs-with-zinit)
+- [Final Verification](#final-verification)
+- [Questions and Feedback](#questions-and-feedback)
+
+***
+
+## Introduction
+
+In this ThreeFold guide, we explore how to set an IPFS node on a micro VM using the ThreeFold Playground.
+
+## Deploy a Micro VM
+
+We start by deploying a micro VM on the ThreeFold Playground.
+
+* Go to the [Threefold Playground](https://playground.grid.tf/#/)
+* Deploy a micro VM (Ubuntu 22.04) with an IPv4 address
+ * IPv4 Address
+ * Minimum vcores: 1vcore
+ * Minimum MB of RAM: 1024MB
+ * Minimum storage: 50GB
+* After deployment, note the VM IPv4 address
+* Connect to the VM via SSH
+ * ```
+ ssh root@VM_IPv4_address
+ ```
+
+## Install the Prerequisites
+
+We install the prerequisites before installing and setting IPFS.
+
+* Update Ubuntu
+ * ```
+ apt update
+ ```
+* Install nano and ufw
+ * ```
+ apt install nano && apt install ufw -y
+ ```
+
+## Set a Firewall
+
+We set a firewall to monitor and control incoming and outgoing network traffic. To do so, we will define predetermined security rules. As a firewall, we will be using [Uncomplicated Firewall](https://wiki.ubuntu.com/UncomplicatedFirewall) (ufw).
+
+For our security rules, we want to allow SSH, HTTP and HTTPS (443 and 8443).
+
+We thus add the following rules:
+
+* Allow SSH (port 22)
+ * ```
+ ufw allow ssh
+ ```
+* Allow port 4001
+ * ```
+ ufw allow 4001
+ ```
+* To enable the firewall, write the following:
+ * ```
+ ufw enable
+ ```
+
+* To see the current security rules, write the following:
+ * ```
+ ufw status verbose
+ ```
+
+You have enabled the firewall with proper security rules for your IPFS deployment.
+
+### Additional Ports
+
+We provided the basic firewall ports for your IPFS instance. There are other more advanced configurations possible.
+
+If you want to access your IPFS node remotely, you can allow **port 5001**. This will allow anyone to access your IPFS node. Make sure that you know what you are doing if you go this route. You should, for example, restrict which external IP address can access port 5001.
+
+If you want to run your deployment as a gateway node, you should allow **port 8080**. Read the IPFS documentation for more information on this.
+
+If you want to run pubsub capabilities, you need to allow **port 8081**. For more information, read the [IPFS documentation](https://blog.ipfs.tech/25-pubsub/).
+
+## Install IPFS
+
+We install the [IPFS Kubo binary](https://docs.ipfs.tech/install/command-line/#install-official-binary-distributions).
+
+* Download the binary
+ * ```
+ wget https://dist.ipfs.tech/kubo/v0.24.0/kubo_v0.24.0_linux-amd64.tar.gz
+ ```
+* Unzip the file
+ * ```
+ tar -xvzf kubo_v0.24.0_linux-amd64.tar.gz
+ ```
+* Change directory
+ * ```
+ cd kubo
+ ```
+* Run the install script
+ * ```
+ bash install.sh
+ ```
+* Verify that IPFS Kubo is properly installed
+ * ```
+ ipfs --version
+ ```
+
+## Set IPFS
+
+We initialize IPFS and run the IPFS daemon.
+
+* Initialize IPFS
+ * ```
+ ipfs init --profile server
+ ```
+* Increase the storage capacity (optional)
+ * ```
+ ipfs config Datastore.StorageMax 30GB
+ ```
+* Run the IPFS daemon
+ * ```
+ ipfs daemon
+ ```
+
+## Set IPFS with zinit
+
+We set the IPFS daemon with zinit. This will make sure that the IPFS daemon starts at each VM reboot or if it stops functioning momentarily.
+
+* Create the yaml file
+ * ```
+ nano /etc/zinit/ipfs.yaml
+ ```
+* Set the execution command
+ * ```
+ exec: /usr/local/bin/ipfs daemon
+ ```
+* Run the IPFS daemon with the zinit monitor command
+ * ```
+ zinit monitor ipfs
+ ```
+* Verify that the IPFS daemon is running
+ * ```
+ ipfs swarm peers
+ ```
+
+## Final Verification
+
+We reboot and reconnect to the VM and verify that IPFS is properly running as a final verification.
+
+* Reboot the VM
+ * ```
+ reboot -f
+ ```
+* Reconnect to the VM and verify that the IPFS daemon is running
+ * ```
+ ipfs swarm peers
+ ```
+
+## Questions and Feedback
+
+If you have any questions or feedback, please let us know by either writing a post on the [ThreeFold Forum](https://forum.threefold.io/), or by chatting with us on the [TF Grid Tester Community](https://t.me/threefoldtesting) Telegram channel.
\ No newline at end of file
diff --git a/collections/documentation/system_administrators/advanced/ipfs/ipfs_toc.md b/collections/documentation/system_administrators/advanced/ipfs/ipfs_toc.md
new file mode 100644
index 0000000..15d9c4a
--- /dev/null
+++ b/collections/documentation/system_administrators/advanced/ipfs/ipfs_toc.md
@@ -0,0 +1,6 @@
+IPFS and ThreeFold
+
+Table of Contents
+
+- [IPFS on a Full VM](./ipfs_fullvm.md)
+- [IPFS on a Micro VM](./ipfs_microvm.md)
\ No newline at end of file
diff --git a/collections/documentation/system_administrators/advanced/list_public_ips.md b/collections/documentation/system_administrators/advanced/list_public_ips.md
new file mode 100644
index 0000000..80fb72a
--- /dev/null
+++ b/collections/documentation/system_administrators/advanced/list_public_ips.md
@@ -0,0 +1,22 @@
+ Listing Public IPs
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Example](#example)
+
+***
+
+## Introduction
+
+Listing public IPs can be done by asking graphQL for all IPs that has `contractId = 0`
+
+## Example
+
+```graphql
+query MyQuery {
+ publicIps(where: {contractId_eq: 0}) {
+ ip
+ }
+}
+```
diff --git a/collections/documentation/system_administrators/advanced/token_transfer_keygenerator.md b/collections/documentation/system_administrators/advanced/token_transfer_keygenerator.md
new file mode 100644
index 0000000..38c9ce1
--- /dev/null
+++ b/collections/documentation/system_administrators/advanced/token_transfer_keygenerator.md
@@ -0,0 +1,88 @@
+
+ Transfer TFT Between Networks by Using the Keygenerator
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Prerequisites](#prerequisites)
+ - [Keypair](#keypair)
+ - [Stellar to TFChain](#stellar-to-tfchain)
+ - [Alternative Transfer to TF Chain](#alternative-transfer-to-tf-chain)
+- [TFChain to Stellar](#tfchain-to-stellar)
+
+***
+
+## Introduction
+
+Using this method, only transfer is possible between accounts that are generated in the same manner and that are yours. Please find the keygen tooling for it below.
+
+## Prerequisites
+
+### Keypair
+
+- ed25519 keypair
+- Go installed on your local computer
+
+Create a keypair with the following tool:
+
+```sh
+go build .
+./keygen
+```
+
+### Stellar to TFChain
+
+Create a Stellar wallet from the key that you generated.
+Transfer the TFT from your wallet to the bridge address. A deposit fee of 1 TFT will be taken, so make sure you send a larger amount as 1 TFT.
+
+Bridge addresses :
+
+- On Mainnet: `GBNOTAYUMXVO5QDYWYO2SOCOYIJ3XFIP65GKOQN7H65ZZSO6BK4SLWSC` on [Stellar Mainnet](https://stellar.expert/explorer/public).
+- On testnet: `GA2CWNBUHX7NZ3B5GR4I23FMU7VY5RPA77IUJTIXTTTGKYSKDSV6LUA4` on [Stellar MAINnet](https://stellar.expert/explorer/public)
+
+The amount deposited on TF Chain minus 1 TFT will be transferred over the bridge to the TFChain account.
+
+Effect will be the following :
+
+- Transferred TFTs from Stellar will be sent to a Stellar vault account representing all tokens on TFChain
+- TFTs will be minted on the TFChain for the transferred amount
+
+### Alternative Transfer to TF Chain
+
+We also enabled deposits to TF Grid objects. Following objects can be deposited to:
+
+- Twin
+- Farm
+- Node
+- Entity
+
+To deposit to any of these objects, a memo text in format `object_objectID` must be passed on the deposit to the bridge wallet. Example: `twin_1`.
+
+To deposit to a TF Grid object, this object **must** exists. If the object is not found on chain, a refund is issued.
+
+## TFChain to Stellar
+
+Create a TFChain account from the key that you generated. (TF Chain raw seed).
+Browse to :
+
+- For mainnet:
+- For testnet:
+- For Devnet: https://polkadot.js.org/apps/?rpc=wss%3A%2F%2Ftfchain.dev.grid.tf#/accounts
+
+-> Add Account -> Click on mnemonic and select `Raw Seed` -> Paste raw TF Chain seed.
+
+Select `Advanced creation options` -> Change `keypair crypto type` to `Edwards (ed25519)`. Click `I have saved my mnemonic seed safely` and proceed.
+
+Choose a name and password and proceed.
+
+Browse to the [extrinsics](https://polkadot.js.org/apps/?rpc=wss%3A%2F%2Ftfchain.test.grid.tf#/extrinsics) , select tftBridgeModule and extrinsic: `swap_to_stellar`. Provide your Bridge substrate address and the amount to transfer. Sign using your password.
+Again, a withdrawfee of 1 TFT will be taken, so make sure you send an amount larger than 1 TFT.
+
+The amount withdrawn from TFChain will be sent to your Stellar wallet.
+
+Behind the scenes, following will happen:
+
+- Transferred TFTs from Stellar will be sent from the Stellar vault account to the user's Stellar account
+- TFTs will be burned on the TFChain for the transferred amount
+
+Example: ![swap_to_stellar](img/swap_to_stellar.png ':size=400')
diff --git a/collections/documentation/system_administrators/computer_it_basics/cli_scripts_basics.md b/collections/documentation/system_administrators/computer_it_basics/cli_scripts_basics.md
new file mode 100644
index 0000000..dec08b2
--- /dev/null
+++ b/collections/documentation/system_administrators/computer_it_basics/cli_scripts_basics.md
@@ -0,0 +1,1138 @@
+
+ CLI and Scripts Basics
+
+ Table of Contents
+
+- [Introduction](#introduction)
+- [Basic Commands](#basic-commands)
+ - [Update and upgrade packages](#update-and-upgrade-packages)
+ - [Test the network connectivity of a domain or an IP address with ping](#test-the-network-connectivity-of-a-domain-or-an-ip-address-with-ping)
+ - [Install Go](#install-go)
+ - [Install Brew](#install-brew)
+ - [Brew basic commands](#brew-basic-commands)
+ - [Install Terraform with Brew](#install-terraform-with-brew)
+ - [Yarn basic commands](#yarn-basic-commands)
+ - [Set default terminal](#set-default-terminal)
+ - [See the current path](#see-the-current-path)
+ - [List hidden files](#list-hidden-files)
+ - [Display the content of a directory](#display-the-content-of-a-directory)
+ - [Vim modes and basic commands](#vim-modes-and-basic-commands)
+ - [Check the listening ports using netstat](#check-the-listening-ports-using-netstat)
+ - [See the disk usage of different folders](#see-the-disk-usage-of-different-folders)
+ - [Verify the application version](#verify-the-application-version)
+ - [Find the path of a file with only the file name](#find-the-path-of-a-file-with-only-the-file-name)
+ - [Become the superuser (su) on Linux](#become-the-superuser-su-on-linux)
+ - [Exit a session](#exit-a-session)
+ - [Know the current user](#know-the-current-user)
+ - [Set the path of a package](#set-the-path-of-a-package)
+ - [See the current path](#see-the-current-path-1)
+ - [Find the current shell](#find-the-current-shell)
+ - [SSH into Remote Server](#ssh-into-remote-server)
+ - [Replace a string by another string in a text file](#replace-a-string-by-another-string-in-a-text-file)
+ - [Replace extensions of files in a folder](#replace-extensions-of-files-in-a-folder)
+ - [Remove extension of files in a folder](#remove-extension-of-files-in-a-folder)
+ - [See the current date and time on Linux](#see-the-current-date-and-time-on-linux)
+ - [Special variables in Bash Shell](#special-variables-in-bash-shell)
+ - [Gather DNS information of a website](#gather-dns-information-of-a-website)
+ - [Partition and mount a disk](#partition-and-mount-a-disk)
+- [Encryption](#encryption)
+ - [Encrypt files with Gocryptfs](#encrypt-files-with-gocryptfs)
+ - [Encrypt files with Veracrypt](#encrypt-files-with-veracrypt)
+- [Network-related Commands](#network-related-commands)
+ - [See the network connections and ports](#see-the-network-connections-and-ports)
+ - [See identity and info of IP address](#see-identity-and-info-of-ip-address)
+ - [ip basic commands](#ip-basic-commands)
+ - [Display socket statistics](#display-socket-statistics)
+ - [Query or control network driver and hardware settings](#query-or-control-network-driver-and-hardware-settings)
+ - [See if ethernet port is active](#see-if-ethernet-port-is-active)
+ - [Add IP address to hardware port (ethernet)](#add-ip-address-to-hardware-port-ethernet)
+ - [Private IP address range](#private-ip-address-range)
+ - [Set IP Address manually](#set-ip-address-manually)
+- [Basic Scripts](#basic-scripts)
+ - [Run a script with arguments](#run-a-script-with-arguments)
+ - [Print all arguments](#print-all-arguments)
+ - [Iterate over arguments](#iterate-over-arguments)
+ - [Count lines in files given as arguments](#count-lines-in-files-given-as-arguments)
+ - [Find path of a file](#find-path-of-a-file)
+ - [Print how many arguments are passed in a script](#print-how-many-arguments-are-passed-in-a-script)
+- [Linux](#linux)
+ - [Install Terraform](#install-terraform)
+- [MAC](#mac)
+ - [Enable remote login on MAC](#enable-remote-login-on-mac)
+ - [Find Other storage on MAC](#find-other-storage-on-mac)
+ - [Sort files by size and extension on MAC](#sort-files-by-size-and-extension-on-mac)
+- [Windows](#windows)
+ - [Install Chocolatey](#install-chocolatey)
+ - [Install Terraform with Chocolatey](#install-terraform-with-chocolatey)
+ - [Find the product key](#find-the-product-key)
+ - [Find Windows license type](#find-windows-license-type)
+- [References](#references)
+
+***
+
+## Introduction
+
+We present here a quick guide on different command-line interface (CLI) commands as well as some basic scripts.
+
+The main goal of this guide is to demonstrate that having some core understanding of CLI and scripts can drastically increase efficiency and speed when it comes to deploying and managing workloads on the TFGrid.
+
+## Basic Commands
+
+### Update and upgrade packages
+
+The command **update** ensures that you have access to the latest versions of packages available.
+
+```
+sudo apt update
+```
+
+The command **upgrade** downloads and installs the updates for each outdated package and dependency on your system.
+
+```
+sudo apt upgrade
+```
+
+
+
+### Test the network connectivity of a domain or an IP address with ping
+
+To test the network connectivity of a domain or an IP address, you can use `ping` on Linux, MAC and Windows:
+
+* Template
+ ```
+ ping
+ ```
+* Example
+ ```
+ ping threefold.io
+ ```
+
+On Windows, by default, the command will send 4 packets. On MAC and Linux, it will keep on sending packets, so you will need to press `Ctrl-C` to stop the command from running.
+
+You can also set a number of counts with `-c` on Linux and MAC and `-n` on Windows.
+
+* Send a given number of packets on Linux and MAC (e.g 5 packets)
+ ```
+ ping -c 5 threefold.io
+ ```
+* Send a given number of packets on Windows (e.g 5 packets)
+ ```
+ ping -n 5 threefold.io
+ ```
+
+***
+
+### Install Go
+
+Here are the steps to install [Go](https://go.dev/).
+
+* Install go
+ * ```
+ sudo apt install golang-go
+ ```
+* Verify that go is properly installed
+ * ```
+ go version
+ ```
+
+
+
+### Install Brew
+
+Follow those steps to install [Brew](https://brew.sh/)
+
+* Installation command from Brew:
+ * ```
+ /bin/bash -c "$(curl -fsSL https://raw.githubusercontent.com/Homebrew/install/HEAD/install.sh)"
+ ```
+* Add the path to the **.profile** directory. Replace by your username.
+ * ```
+ echo 'eval "$(/home/linuxbrew/.linuxbrew/bin/brew shellenv)"' >> /home//.profile
+ ```
+* Evaluation the following:
+ * ```
+ eval "$(/home/linuxbrew/.linuxbrew/bin/brew shellenv)"
+ ```
+* Verify the installation
+ * ```
+ brew doctor
+ ```
+
+
+
+### Brew basic commands
+
+* To update brew in general:
+ * ```
+ brew update
+ ```
+* To update a specific package:
+ * ```
+ brew update
+ ```
+* To install a package:
+ * ```
+ brew install
+ ```
+* To uninstall a package:
+ * ```
+ brew uninstall
+ ```
+* To search a package:
+ * ```
+ brew search
+ ```
+* [Uninstall Brew](https://github.com/homebrew/install#uninstall-homebrew)
+ * ```
+ /bin/bash -c "$(curl -fsSL https://raw.githubusercontent.com/Homebrew/install/HEAD/uninstall.sh)"
+ ```
+
+
+
+### Install Terraform with Brew
+
+Installing Terraform with Brew is very simple by following the [Terraform documentation](https://developer.hashicorp.com/terraform/downloads).
+
+* Compile HashiCorp software on Homebrew's infrastructure
+ * ```
+ brew tap hashicorp/tap
+ ```
+* Install Terraform
+ * ```
+ brew install hashicorp/tap/terraform
+ ```
+
+
+
+### Yarn basic commands
+
+* Add a package
+ * ```
+ yarn add
+ ```
+* Initialize the development of a package
+ * ```
+ yarn init
+ ```
+* Install all the dependencies in the **package.json** file
+ * ```
+ yarn install
+ ```
+* Publish a package to a package manager
+ * ```
+ yarn publish
+ ```
+* Remove unused package from the current package
+ * ```
+ yarn remove
+ ```
+* Clean the cache
+ * ```
+ yarn cache clean
+ ```
+
+
+
+### Set default terminal
+
+```
+update-alternatives --config x-terminal-emulator
+```
+
+### See the current path
+
+```
+pwd
+```
+
+
+
+### List hidden files
+
+```
+ls -ld .?*
+```
+
+
+
+### Display the content of a directory
+
+You can use **tree** to display the files and organization of a directory:
+
+* General command
+ * ```
+ tree
+ ```
+* View hidden files
+ * ```
+ tree -a
+ ```
+
+
+
+### Vim modes and basic commands
+
+[Vim](https://www.vim.org/) is a free and open-source, screen-based text editor program.
+
+With Vim, you can use two modes.
+
+* Insert mode - normal text editor
+ * Press **i**
+* Command mode - commands to the editor
+ * Press **ESC**
+
+Here are some basic commands:
+
+* Delete characters
+ * **x**
+* Undo last command
+ * **u**
+* Undo the whole line
+ * **U**
+* Go to the end of line
+ * **A**
+* Save and exit
+ * **:wq**
+* Discard all changes
+ * **:q!**
+* Move cursor to the start of the line
+ * **0**
+* Delete the current word
+ * **dw**
+* Delete the current line
+ * **dd**
+
+
+
+### Check the listening ports using netstat
+
+Use the command:
+
+```
+netstat
+```
+
+
+
+
+### See the disk usage of different folders
+
+```
+du -sh *
+```
+
+
+
+
+### Verify the application version
+
+```
+which
+```
+
+
+
+### Find the path of a file with only the file name
+
+On MAC and Linux, you can use **coreutils** and **realpath** from Brew:
+
+* ```
+ brew install coreutils
+ ```
+* ```
+ realpath file_name
+ ```
+
+
+
+### Become the superuser (su) on Linux
+
+You can use either command:
+
+* Option 1
+ * ```
+ sudo -i
+ ```
+* Option 2
+ * ```
+ sudo -s
+ ```
+
+
+
+### Exit a session
+
+You can use either command depending on your shell:
+
+* ```
+ exit
+ ```
+* ```
+ logout
+ ```
+
+
+
+### Know the current user
+
+You can use the following command:
+
+* ```
+ whoami
+ ```
+
+
+
+### See the path of a package
+
+To see the path of a package, you can use the following command:
+
+* ```
+ whereis
+ ```
+
+
+
+### Set the path of a package
+
+```
+export PATH=$PATH:/snap/bin
+
+```
+
+
+
+
+### See the current path
+
+```
+pwd
+```
+
+
+
+### Find the current shell
+
+* Compact version
+ * ```
+ echo $SHELL
+ ```
+* Detailed version
+ * ```
+ ls -l /proc/$$/exe
+ ```
+
+
+
+### SSH into Remote Server
+
+* Create SSH key pair
+ * ```
+ ssh-keygen
+ ```
+* Install openssh-client on the local computer*
+ * ```
+ sudo apt install openssh-client
+ ```
+* Install openssh-server on the remote computer*
+ * ```
+ sudo apt install openssh-server
+ ```
+* Copy public key
+ * ```
+ cat ~/.ssh/id_rsa.pub
+ ```
+* Create the ssh directory on the remote computer
+ * ```
+ mkdir ~/.ssh
+ ```
+* Add public key in the file **authorized_keys** on the remote computer
+ * ```
+ nano ~/.ssh/authorized_keys
+ ```
+* Check openssh-server status
+ * ```
+ sudo service ssh status
+ ```
+* SSH into the remote machine
+ * ```
+ ssh @
+ ```
+
+\*Note: For MAC, you can install **openssh-server** and **openssh-client** with Brew: **brew install openssh-server** and **brew install openssh-client**.
+
+To enable remote login on a MAC, [read this section](#enable-remote-login-on-mac).
+
+
+
+### Replace a string by another string in a text file
+
+* Replace one string by another (e.g. **old_string**, **new_string**)
+ * ```
+ sed -i 's/old_string/new_string/g' /
+ ```
+* Use environment variables (double quotes)
+ * ```
+ sed -i "s/old_string/$env_variable/g" /
+ ```
+
+
+
+### Replace extensions of files in a folder
+
+Replace **ext1** and **ext2** by the extensions in question.
+
+```
+find ./ -depth -name "*.ext1" -exec sh -c 'mv "$1" "${1%.ext1}.ext2"' _ {} \;
+```
+
+
+
+### Remove extension of files in a folder
+
+Replace **ext** with the extension in question.
+
+```bash
+for file in *.ext; do
+ mv -- "$file" "${file%%.ext}"
+done
+```
+
+
+
+### See the current date and time on Linux
+
+```
+date
+```
+
+
+
+### Special variables in Bash Shell
+
+| Special Variables | Descriptions |
+| ---------------- | ----------------------------------------------- |
+| $0 | Name of the bash script |
+| $1, $2...$n | Bash script arguments |
+| $$ | Process id of the current shell |
+| $* | String containing every command-line argument |
+| $# | Total number of arguments passed to the script |
+| $@ | Value of all the arguments passed to the script |
+| $? | Exit status of the last executed command |
+| $! | Process id of the last executed command |
+| $- | Print current set of option in current shell |
+
+
+
+### Gather DNS information of a website
+
+You can use [Dig](https://man.archlinux.org/man/dig.1) to gather DNS information of a website
+
+* Template
+ * ```
+ dig
+ ```
+* Example
+ * ```
+ dig threefold.io
+ ```
+
+You can also use online tools such as [DNS Checker](https://dnschecker.org/).
+
+
+
+### Partition and mount a disk
+
+We present one of many ways to partition and mount a disk.
+
+* Create partition with [gparted](https://gparted.org/)
+ * ```
+ sudo gparted
+ ```
+* Find the disk you want to mount (e.g. **sdb**)
+ * ```
+ sudo fdisk -l
+ ```
+* Create a directory to mount the disk to
+ * ```
+ sudo mkdir /mnt/disk
+ ```
+* Open fstab
+ * ```
+ sudo nano /etc/fstab
+ ```
+* Append the following to the fstab with the proper disk path (e.g. **/dev/sdb**) and mount point (e.g. **/mnt/disk**)
+ * ```
+ /dev/sdb /mnt/disk ext4 defaults 0 0
+ ```
+* Mount the disk
+ * ```
+ sudo mount /mnt/disk
+ ```
+* Add permissions (as needed)
+ * ```
+ sudo chmod -R 0777 /mnt/disk
+ ```
+
+
+
+## Encryption
+
+### Encrypt files with Gocryptfs
+
+You can use [gocryptfs](https://github.com/rfjakob/gocryptfs) to encrypt files.
+
+* Install gocryptfs
+ * ```
+ apt install gocryptfs
+ ```
+* Create a vault directory (e.g. **vaultdir**) and a mount directory (e.g. **mountdir**)
+ * ```
+ mkdir vaultdir mountdir
+ ```
+* Initiate the vault
+ * ```
+ gocryptfs -init vaultdir
+ ```
+* Mount the mount directory with the vault
+ * ```
+ gocryptfs vaultdir mountdir
+ ```
+* You can now create files in the folder. For example:
+ * ```
+ touch mountdir/test.txt
+ ```
+* The new file **test.txt** is now encrypted in the vault
+ * ```
+ ls vaultdir
+ ```
+* To unmount the mountedvault folder:
+ * Option 1
+ * ```
+ fusermount -u mountdir
+ ```
+ * Option 2
+ * ```
+ rmdir mountdir
+ ```
+
+
+### Encrypt files with Veracrypt
+
+To encrypt files, you can use [Veracrypt](https://www.veracrypt.fr/en/Home.html). Let's see how to download and install Veracrypt.
+
+* Veracrypt GUI
+ * Download the package
+ * ```
+ wget https://launchpad.net/veracrypt/trunk/1.25.9/+download/veracrypt-1.25.9-Ubuntu-22.04-amd64.deb
+ ```
+ * Install the package
+ * ```
+ dpkg -i ./veracrypt-1.25.9-Ubuntu-22.04-amd64.deb
+ ```
+* Veracrypt console only
+ * Download the package
+ * ```
+ wget https://launchpad.net/veracrypt/trunk/1.25.9/+download/veracrypt-console-1.25.9-Ubuntu-22.04-amd64.deb
+ ```
+ * Install the package
+ * ```
+ dpkg -i ./veracrypt-console-1.25.9-Ubuntu-22.04-amd64.deb
+ ```
+
+You can visit [Veracrypt download page](https://www.veracrypt.fr/en/Downloads.html) to get the newest releases.
+
+* To run Veracrypt
+ * ```
+ veracrypt
+ ```
+* Veracrypt documentation is very complete. To begin using the application, visit the [Beginner's Tutorial](https://www.veracrypt.fr/en/Beginner%27s%20Tutorial.html).
+
+
+
+## Network-related Commands
+
+### See the network connections and ports
+
+ifconfig
+
+
+
+### See identity and info of IP address
+
+* See abuses related to an IP address:
+ * ```
+ https://www.abuseipdb.com/check/
+ ```
+* See general information of an IP address:
+ * ```
+ https://www.whois.com/whois/
+ ```
+
+
+
+### ip basic commands
+
+* Manage and display the state of all network
+ * ```
+ ip link
+ ```
+* Display IP Addresses and property information (abbreviation of address)
+ * ```
+ ip addr
+ ```
+* Display and alter the routing table
+ * ```
+ ip route
+ ```
+* Manage and display multicast IP addresses
+ * ```
+ ip maddr
+ ```
+* Show neighbour object
+ * ```
+ ip neigh
+ ```
+* Display a list of commands and arguments for
+each subcommand
+ * ```
+ ip help
+ ```
+* Add an address
+ * Template
+ * ```
+ ip addr add
+ ```
+ * Example: set IP address to device **enp0**
+ * ```
+ ip addr add 192.168.3.4/24 dev enp0
+ ```
+* Delete an address
+ * Template
+ * ```
+ ip addr del
+ ```
+ * Example: set IP address to device **enp0**
+ * ```
+ ip addr del 192.168.3.4/24 dev enp0
+ ```
+* Alter the status of an interface
+ * Template
+ * ```
+ ip link set
+ ```
+ * Example 1: Bring interface online (here device **em2**)
+ * ```
+ ip link set em2 up
+ ```
+ * Example 2: Bring interface offline (here device **em2**)
+ * ```
+ ip link set em2 down
+ ```
+* Add a multicast address
+ * Template
+ * ```
+ ip maddr add
+ ```
+ * Example : set IP address to device **em2**
+ * ```
+ ip maddr add 33:32:00:00:00:01 dev em2
+ ```
+* Delete a multicast address
+ * Template
+ * ```
+ ip maddr del
+ ```
+ * Example: set IP address to device **em2**
+ * ```
+ ip maddr del 33:32:00:00:00:01 dev em2
+ ```
+* Add a routing table entry
+ * Template
+ * ```
+ ip route add
+ ```
+ * Example 1: Add a default route (for all addresses) via a local gateway
+ * ```
+ ip route add default via 192.168.1.1 dev em1
+ ```
+ * Example 2: Add a route to 192.168.3.0/24 via the gateway at 192.168.3.2
+ * ```
+ ip route add 192.168.3.0/24 via 192.168.3.2
+ ```
+ * Example 3: Add a route to 192.168.1.0/24 that can be reached on
+device em1
+ * ```
+ ip route add 192.168.1.0/24 dev em1
+ ```
+* Delete a routing table entry
+ * Template
+ * ```
+ ip route delete
+ ```
+ * Example: Delete the route for 192.168.1.0/24 via the gateway at
+192.168.1.1
+ * ```
+ ip route delete 192.168.1.0/24 via 192.168.1.1
+ ```
+* Replace, or add, a route
+ * Template
+ * ```
+ ip route replace
+ ```
+ * Example: Replace the defined route for 192.168.1.0/24 to use
+device em1
+ * ```
+ ip route replace 192.168.1.0/24 dev em1
+ ```
+* Display the route an address will take
+ * Template
+ * ```
+ ip route get
+ ```
+ * Example: Display the route taken for IP 192.168.18.25
+ * ```
+ ip route replace 192.168.18.25/24 dev enp0
+ ```
+
+
+
+References: https://www.commandlinux.com/man-page/man8/ip.8.html
+
+
+
+### Display socket statistics
+
+* Show all sockets
+ * ```
+ ss -a
+ ```
+* Show detailed socket information
+ * ```
+ ss -e
+ ```
+* Show timer information
+ * ```
+ ss -o
+ ```
+* Do not resolve address
+ * ```
+ ss -n
+ ```
+* Show process using the socket
+ * ```
+ ss -p
+ ```
+
+Note: You can combine parameters, e.g. **ss -aeo**.
+
+References: https://www.commandlinux.com/man-page/man8/ss.8.html
+
+
+
+### Query or control network driver and hardware settings
+
+* Display ring buffer for a device (e.g. **eth0**)
+ * ```
+ ethtool -g eth0
+ ```
+* Display driver information for a device (e.g. **eth0**)
+ * ```
+ ethtool -i eth0
+ ```
+* Identify eth0 by sight, e.g. by causing LEDs to blink on the network port
+ * ```
+ ethtool -p eth0
+ ```
+* Display network and driver statistics for a device (e.g. **eth0**)
+ * ```
+ ethtool -S eth0
+ ```
+
+References: https://man.archlinux.org/man/ethtool.8.en
+
+
+
+### See if ethernet port is active
+
+Replace with the proper device:
+
+```
+cat /sys/class/net//carrier
+```
+
+
+
+### Add IP address to hardware port (ethernet)
+
+* Find ethernet port ID on both computers
+ * ```
+ ip a
+ ```
+* Add IP address (DHCO or static)
+ * Computer 1
+ * ```
+ ip addr add /24 dev
+ ```
+ * Computer 2
+ * ```
+ ip addr add /24 dev
+ ```
+
+* [Ping](#test-the-network-connectivity-of-a-domain-or-an-ip-address-with-ping) the address to confirm connection
+ * ```
+ ping
+ ```
+
+To set and view the address for either DHCP or static, go to **Networks** then **Details**.
+
+
+
+### Private IP address range
+
+The private IP range is the following:
+
+* 10.0.0.0–10.255.255.255
+* 172.16.0.0–172.31.255.255
+* 192.168.0.0–192.168.255.255
+
+
+
+### Set IP Address manually
+
+You can use the following template when you set an IP address manually:
+
+* Address
+ *
+* Netmask
+ * 255.255.255.0
+* Gateway
+ * optional
+
+
+
+## Basic Scripts
+
+### Run a script with arguments
+
+You can use the following template to add arguments when running a script:
+
+* Option 1
+ * ```
+ ./example_script.sh arg1 arg2
+ ```
+* Option 2
+ * ```
+ sh example_script.sh "arg1" "arg2"
+ ```
+
+### Print all arguments
+
+* Write a script
+ * File: `example_script.sh`
+ * ```bash
+ #!/bin/sh
+ echo $@
+ ```
+* Give permissions
+ * ```bash
+ chmod +x ./example_script.sh
+ ```
+* Run the script with arguments
+ * ```bash
+ sh example_script.sh arg1 arg2
+ ```
+
+
+### Iterate over arguments
+
+* Write the script
+ * ```bash
+ # iterate_script.sh
+ #!/bin/bash
+ for i; do
+ echo $i
+ done
+ ```
+* Give permissions
+ * ```
+ chmod +x ./iterate_script.sh
+ ```
+* Run the script with arguments
+ * ```
+ sh iterate_script.sh arg1 arg2
+ ```
+
+* The following script is equivalent
+ * ```bash
+ # iterate_script.sh
+ #/bin/bash
+ for i in $*; do
+ echo $i
+ done
+ ```
+
+
+
+### Count lines in files given as arguments
+
+* Write the script
+ * ```bash
+ # count_lines.sh
+ #!/bin/bash
+ for i in $*; do
+ nlines=$(wc -l < $i)
+ echo "There are $nlines lines in $i"
+ done
+ ```
+* Give permissions
+ * ```
+ chmod +x ./count_lines.sh
+ ```
+* Run the script with arguments (files). Here we use the script itself as an example.
+ * ```
+ sh count_lines.sh count_lines.sh
+ ```
+
+
+
+### Find path of a file
+
+* Write the script
+ * ```bash
+ # find.sh
+ #!/bin/bash
+
+ find / -iname $1 2> /dev/null
+ ```
+* Run the script
+ * ```
+ sh find.sh
+ ```
+
+
+
+### Print how many arguments are passed in a script
+
+* Write the script
+ * ```bash
+ # print_qty_args.sh
+ #!/bin/bash
+ echo This script was passed $# arguments
+ ```
+* Run the script
+ * ```
+ sh print_qty_args.sh
+ ```
+
+
+## Linux
+
+### Install Terraform
+
+Here are the steps to install Terraform on Linux based on the [Terraform documentation](https://developer.hashicorp.com/terraform/downloads).
+
+```
+wget -O- https://apt.releases.hashicorp.com/gpg | sudo gpg --dearmor -o /usr/share/keyrings/hashicorp-archive-keyring.gpg
+```
+```
+echo "deb [signed-by=/usr/share/keyrings/hashicorp-archive-keyring.gpg] https://apt.releases.hashicorp.com $(lsb_release -cs) main" | sudo tee /etc/apt/sources.list.d/hashicorp.list
+```
+```
+sudo apt update && sudo apt install terraform
+```
+
+Note that the Terraform documentation also covers other methods to install Terraform on Linux.
+
+## MAC
+
+### Enable remote login on MAC
+
+* Option 1:
+ * Use the following command line:
+ * ```
+ systemsetup -setremotelogin on
+ ```
+* Option 2
+ * Use **System Preferences**
+ * Go to **System Preferences** -> **Sharing** -> **Enable Remote Login**.
+
+
+
+### Find Other storage on MAC
+
+* Open **Finder** \> **Go** \> **Go to Folder**
+* Paste this path
+ * ```
+ ~/Library/Caches
+ ```
+
+
+
+### Sort files by size and extension on MAC
+
+* From your desktop, press **Command-F**.
+* Click **This Mac**.
+* Click the first dropdown menu field and select **Other**.
+* From the **Search Attributes** window
+ * tick **File Size** and **File Extension**.
+
+
+
+## Windows
+
+### Install Chocolatey
+
+To install Chocolatey on Windows, we follow the [official Chocolatey website](https://chocolatey.org/install) instructions.
+
+* Run PowerShell as Administrator
+* Check if **Get-ExecutionPolicy** is restricted
+ * ```
+ Get-ExecutionPolicy
+ ```
+ * If it is restricted, run the following command:
+ * ```
+ Set-ExecutionPolicy AllSigned
+ ```
+* Install Chocolatey
+ * ```
+ Set-ExecutionPolicy Bypass -Scope Process -Force; [System.Net.ServicePointManager]::SecurityProtocol = [System.Net.ServicePointManager]::SecurityProtocol -bor 3072; iex ((New-Object System.Net.WebClient).DownloadString('https://community.chocolatey.org/install.ps1'))
+ ```
+* Note: You might need to restart PowerShell to use Chocolatey
+
+
+
+### Install Terraform with Chocolatey
+
+Once you've installed Chocolatey on Windows, installing Terraform is as simple as can be:
+
+* Install Terraform with Chocolatey
+ * ```
+ choco install terraform
+ ```
+
+
+
+### Find the product key
+
+Write the following in **Command Prompt** (run as administrator):
+
+```
+wmic path SoftwareLicensingService get OA3xOriginalProductKey
+```
+
+
+
+### Find Windows license type
+
+Write the following in **Command Prompt**:
+
+```
+slmgr /dli
+```
+
+
+
+## References
+
+* GNU Bash Manual - https://www.gnu.org/software/bash/manual/bash.html
\ No newline at end of file
diff --git a/collections/documentation/system_administrators/computer_it_basics/computer_it_basics.md b/collections/documentation/system_administrators/computer_it_basics/computer_it_basics.md
new file mode 100644
index 0000000..c1dfdfd
--- /dev/null
+++ b/collections/documentation/system_administrators/computer_it_basics/computer_it_basics.md
@@ -0,0 +1,15 @@
+ Computer and IT Basics
+
+Welcome to the *Computer and IT Basics* section of the ThreeFold Manual!
+
+In this section, tailored specifically for system administrators, we'll delve into fundamental concepts and tools that form the backbone of managing and securing infrastructure. Whether you're a seasoned sysadmin or just starting your journey, these basics are essential for navigating the intricacies of the ThreeFold Grid.
+
+ Table of Contents
+
+- [CLI and Scripts Basics](./cli_scripts_basics.md)
+- [Docker Basics](./docker_basics.md)
+- [Git and GitHub Basics](./git_github_basics.md)
+- [Firewall Basics](./firewall_basics/firewall_basics.md)
+ - [UFW Basics](./firewall_basics/ufw_basics.md)
+ - [Firewalld Basics](./firewall_basics/firewalld_basics.md)
+- [File Transfer](./file_transfer.md)
\ No newline at end of file
diff --git a/collections/documentation/system_administrators/computer_it_basics/docker_basics.md b/collections/documentation/system_administrators/computer_it_basics/docker_basics.md
new file mode 100644
index 0000000..5a7f297
--- /dev/null
+++ b/collections/documentation/system_administrators/computer_it_basics/docker_basics.md
@@ -0,0 +1,458 @@
+Docker Basic Commands
+
+Table of Contents
+
+- [Introduction](#introduction)
+- [Basic Commands](#basic-commands)
+ - [Install Docker Desktop and Docker Engine](#install-docker-desktop-and-docker-engine)
+ - [Remove completely Docker](#remove-completely-docker)
+ - [List containers](#list-containers)
+ - [Pull an image](#pull-an-image)
+ - [Push an image](#push-an-image)
+ - [Inspect and pull an image with GHCR](#inspect-and-pull-an-image-with-ghcr)
+ - [See a docker image (no download)](#see-a-docker-image-no-download)
+ - [Build a container](#build-a-container)
+ - [List all available docker images](#list-all-available-docker-images)
+ - [Run a container](#run-a-container)
+ - [Run a new command in an existing container](#run-a-new-command-in-an-existing-container)
+ - [Bash shell into container](#bash-shell-into-container)
+ - [Pass arguments with a bash script and a Dockerfile](#pass-arguments-with-a-bash-script-and-a-dockerfile)
+ - [Copy files from a container to the local computer](#copy-files-from-a-container-to-the-local-computer)
+ - [Delete all the containers, images and volumes](#delete-all-the-containers-images-and-volumes)
+ - [Kill all the Docker processes](#kill-all-the-docker-processes)
+ - [Output full logs for all containers](#output-full-logs-for-all-containers)
+- [Resources Usage](#resources-usage)
+ - [Examine containers with size](#examine-containers-with-size)
+ - [Examine disks usage](#examine-disks-usage)
+- [Wasted Resources](#wasted-resources)
+ - [Prune the Docker logs](#prune-the-docker-logs)
+ - [Prune the Docker containers](#prune-the-docker-containers)
+ - [Remove unused and untagged local container images](#remove-unused-and-untagged-local-container-images)
+ - [Clean up and delete all unused container images](#clean-up-and-delete-all-unused-container-images)
+ - [Clean up container images based on a given timeframe](#clean-up-container-images-based-on-a-given-timeframe)
+- [Command Combinations](#command-combinations)
+ - [Kill all running containers](#kill-all-running-containers)
+ - [Stop all running containers](#stop-all-running-containers)
+ - [Delete all stopped containers](#delete-all-stopped-containers)
+ - [Delete all images](#delete-all-images)
+ - [Update and stop a container in a crash-loop](#update-and-stop-a-container-in-a-crash-loop)
+- [References](#references)
+
+***
+
+## Introduction
+
+We present here a quick introduction to Docker. We cover basic commands, as well as command combinations. Understanding the following should give system administrators confidence when it comes to using Docker efficiently.
+
+The following can serve as a quick reference guide when deploying workloads on the ThreeFold Grid and using Docker in general.
+
+We invite the readers to consult the [official Docker documentation](https://docs.docker.com/) for more information.
+
+
+
+## Basic Commands
+
+### Install Docker Desktop and Docker Engine
+
+You can install [Docker Desktop](https://docs.docker.com/get-docker/) and [Docker Engine](https://docs.docker.com/engine/install/) for Linux, MAC and Windows. Follow the official Docker documentation for the details.
+
+Note that the quickest way to install Docker Engine is to use the convenience script:
+
+```
+curl -fsSL https://get.docker.com -o get-docker.sh
+sudo sh get-docker.sh
+```
+
+
+
+### Remove completely Docker
+
+To completely remove docker from your machine, you can follow these steps:
+
+* List the docker packages
+ * ```
+ dpkg -l | grep -i docker
+ ```
+* Purge and autoremove docker
+ * ```
+ apt-get purge -y docker-engine docker docker.io docker-ce docker-ce-cli docker-compose-plugin
+ apt-get autoremove -y --purge docker-engine docker docker.io docker-ce docker-compose-plugin
+ ```
+* Remove the docker files and folders
+ * ```
+ rm -rf /var/lib/docker /etc/docker
+ rm /etc/apparmor.d/docker
+ groupdel docker
+ rm -rf /var/run/docker.sock
+ ```
+
+You can also use the command **whereis docker** to see if any Docker folders and files remain. If so, remove them with
+
+
+
+### List containers
+
+* List only running containers
+ * ```
+ docker ps
+ ```
+* List all containers (running + stopped)
+ * ```
+ docker ps -a
+ ```
+
+
+
+### Pull an image
+
+To pull an image from [Docker Hub](https://hub.docker.com/):
+
+* Pull an image
+ * ```
+ docker pull
+ ```
+* Pull an image with the tag
+ * ```
+ docker pull :tag
+ ```
+* Pull all tags of an image
+ * ```
+ docker pull -a
+ ```
+
+
+
+### Push an image
+
+To pull an image to [Docker Hub](https://hub.docker.com/):
+
+* Push an image
+ * ```
+ docker push
+ ```
+* Push an image with the tag
+ * ```
+ docker push :tag
+ ```
+* Push all tags of an image
+ * ```
+ docker pull -a
+ ```
+
+
+
+### Inspect and pull an image with GHCR
+
+* Inspect the docker image
+ * ```
+ docker inspect ghcr.io//:
+ ```
+* Pull the docker image
+ * ```
+ docker pull ghcr.io//:
+ ```
+
+
+
+### See a docker image (no download)
+
+If you want to see a docker image without downloading the image itself, you can use Quay's [Skopeo tool](https://github.com/containers/skopeo), a command line utility that performs various operations on container images and image repositories.
+
+```
+docker run --rm quay.io/skopeo/stable list-tags docker://ghcr.io//
+```
+
+Make sure to write the proper information for the repository and the image.
+
+To install Skopeo, read [this documentation](https://github.com/containers/skopeo/blob/main/install.md).
+
+
+
+
+### Build a container
+
+Use **docker build** to build a container based on a Dockerfile
+
+* Build a container based on current directory Dockerfile
+ * ```
+ docker build .
+ ```
+* Build a container and store the image with a given name
+ * Template
+ * ```
+ docker build -t ":"
+ ```
+ * Example
+ * ```
+ docker build -t newimage:latest
+ ```
+* Build a docker container without using the cache
+ * ```
+ docker build --no-cache
+ ```
+
+
+
+### List all available docker images
+
+```
+docker images
+```
+
+
+
+### Run a container
+
+To run a container based on an image, use the command **docker run**.
+
+* Run an image
+ * ```
+ docker run
+ ```
+* Run an image in the background (run and detach)
+ * ```
+ docker run -d
+ ```
+* Run an image with CLI input
+ * ```
+ docker run -it
+ ```
+
+You can combine arguments, e.g. **docker run -itd**.
+
+You can also specify the shell, e.g. **docker run -it /bin/bash**
+
+
+
+### Run a new command in an existing container
+
+To run a new command in an existing container, use **docker exec**.
+
+* Execute interactive shell on the container
+ * ```
+ docker exec -it sh
+ ```
+
+
+
+### Bash shell into container
+
+* Bash shell into a container
+ * ```
+ docker exec -i -t /bin/bash
+ ```
+* Bash shell into a container with root
+ * ```
+ docker exec -i -t -u root /bin/bash
+ ```
+
+Note: if bash is not available, you can use `/bin/sh`
+
+
+
+### Pass arguments with a bash script and a Dockerfile
+
+You can do the following to pass arguments with a bash script and a Dockerfile.
+
+```sh
+# script_example.sh
+#!/bin/sh
+
+echo This is the domain: $env_domain
+echo This is the name: $env_name
+echo This is the password: $env_password
+
+```
+* File `Dockerfile`
+
+```Dockerfile
+FROM ubuntu:latest
+
+ARG domain
+
+ARG name
+
+ARG password
+
+ENV env_domain $domain
+
+ENV env_name $name
+
+ENV env_password $password
+
+COPY script_example.sh .
+
+RUN chmod +x /script_example.sh
+
+CMD ["/script_example.sh"]
+```
+
+
+
+### Copy files from a container to the local computer
+
+```
+docker cp :
+```
+
+
+
+### Delete all the containers, images and volumes
+
+* To delete all containers:
+ * ```
+ docker compose rm -f -s -v
+ ```
+
+* To delete all images:
+ * ```
+ docker rmi -f $(docker images -aq)
+ ```
+
+* To delete all volumes:
+ * ```
+ docker volume rm $(docker volume ls -qf dangling=true)
+ ```
+
+* To delete all containers, images and volumes:
+ * ```
+ docker compose rm -f -s -v && docker rmi -f $(docker images -aq) && docker volume rm $(docker volume ls -qf dangling=true)
+ ```
+
+
+
+### Kill all the Docker processes
+
+* To kill all processes:
+ * ```
+ killall Docker && open /Applications/Docker.app
+ ```
+
+
+
+### Output full logs for all containers
+
+The following command output the full logs for all containers in the file **containers.log**:
+
+```
+docker compose logs > containers.log
+```
+
+
+
+## Resources Usage
+
+### Examine containers with size
+
+```
+docker ps -s
+```
+
+
+
+### Examine disks usage
+
+* Basic mode
+ * ```
+ docker system df
+ ```
+* Verbose mode
+ * ```
+ docker system df -v
+ ```
+
+
+
+## Wasted Resources
+
+### Prune the Docker logs
+
+```
+docker system prune
+```
+
+### Prune the Docker containers
+
+You can use the prune function to delete all stopped containers:
+
+```
+docker container prune
+```
+
+### Remove unused and untagged local container images
+
+The following is useful if you want to clean up local filesystem:
+
+```
+docker image prune
+```
+
+### Clean up and delete all unused container images
+
+```
+docker image prune -a
+```
+
+### Clean up container images based on a given timeframe
+
+To clean up container images created X hours ago, you can use the following template (replace with a number):
+
+```
+docker image prune -a --force --filter "until=h"
+```
+
+To clean up container images created before a given date, you can use the following template (replace