Retirar/reactivar el nodo de Hadoop
I can find info regarding the process of decommissioning of a node in hadoop cluster but not what happens to the data in the decommissioned node.
Only info i found is the answer given by Tariq at ¿La clausura de un nodo elimina datos de ese nodo?
My doubt is regarding the recommission of an old node.
If data is lost, it is same as commissioning a new node. Problem solved.
But, if the data is not lost, when we recommission it, the blocks present in this node become redundant since these blocks were already copied during decommission process. This leads to inconsistency in replicas of the blocks.
How does the hadoop framework takes care of this situation?
preguntado el 25 de noviembre de 13 a las 07:11
No es la respuesta que estás buscando? Examinar otras preguntas etiquetadas hadoop hdfs or haz tu propia pregunta.
May be it'll process over replicated blocks on recommission. Check for processOverReplicatedBlocksOnReCommission in este enlace - SSaikia_JtheRocker
so, whenever a node gets added/removed, it might call processOverReplicatedBlock(). Anyway to validate it? - Krishna
May be you can try: hadoop fsck -blocks to get a report - SSaikia_JtheRocker