One embodiment is directed to a method comprising tracking a movement of a first user’s eye using a first head mounted display device. A depth of focus of the first user’s eye is estimated based on the tracked eye movement. A light beam associated with a display object is modified based on the estimated depth of focus such that the display object appears in focus. The modified light beam is projected toward a display lens of the first head mounted display device. The light beam is directed into the first user’s eye using the display lens. A transmission of light from a local environment of the first user is selectively allowed based on at least a selection of an augmented reality mode of the first head mounted display device. A field-of-view image is captured by the first head mounted display device at the local environment of the first user, wherein the field-of-view image corresponds to a field-of-view for each of the first user’s eyes. A set of points is extracted in the captured field-of-view image. At least one of the extracted set of points is tagged to a type of object. A physical object is recognized as belonging to the type of object based on the at least one of the tagged set of points, and at least a portion of virtual world data associated with the recognized physical object is transmitted to a second head mounted display device associated with a second user at a second location.