

























by Pranav ·


























by Pranav ·
Those of you who have read the Ranger’s Apprentice series by John Flanigan or The King’s Ranger Series by A.C. Cobble will know the importance of staying still and observing movement.
The basic idea is that you stay still and watch your surroundings. The thing you are tracking will eventually move which is when you notice it and take whatever action you have to.
I have wanted to try this with the vOICe which I did today while eating lunch at the Select City Walk mall. Sacheta and I were seated on a bench and I had the vOICe running. I was using my Vuzix M4000 monocular device and watching the scene in front of us.
I could sense how it changed as people moved about and once I managed to sense a slow moving open buss for children. It was like a train. I could hear it but it was fun to see it as well.
This skill also comes in handy while standing in a queue or even in an isle in a shop. I spend about 3 to 4 frames building a mental map of what is in front of me. That means that the soundscapes have panned about 4 times. I then watch for changes. Focus on far objects that is the background. Is something coming nearer? Is the cluster of things reducing? A reduction of things and more space usually means that people have moved out of that area and I am free to move. If a cluster of objects comes closer, it may be time to find another location or just move a little backwards since people want to cross.
I do need to keep revising my mental map. A cluster of people may be sitting or standing for a while which means they become a part of the scene. This is context dependent.
As with most things, this skill comes with practice and it is a good idea to think about what you want to do and then start implementing with the vOICe.
You do have change detection mode in the vOICe that does this automatically but the thing to note with that feature is that it will trigger if you move your head so it stick to the default settings, be persistent and keep practicing until watching for changes becomes second nature.
by Pranav ·
I was walking around in the gated community where I live. Checkout this interesting image I took. It primarily shows different buildings. I was on the upper deck of a swimmingpool area where there was a pond and was surrounded by buildings.
This layer’s soundscape is distinct and is potentially handy to know when you are looking at a wall having multiple breaks which could mean multiple buildings. Also see the pond and hut.

by Pranav ·
Video conferences have proliferated especially after COVID. These days, you are expected to be on camera. This applies even if you are blind and are dealing with sighted people.
I do not have a problem with this per se since it represents equality in a way. Ensuring your face is centered is easy enough with a variety of AI tools available. The Jaws for windows screen reader has this feature built-in while NVDA has an add-on that does the same thing. I believe there also was a standalone program that did this.
However, none of these programs tell you how you are looking. Yes, you can use the plethora of AI apps to check this using your smart phone but remember the context. You are seated on your chair and need to see the picture in that situation.
The vOICe makes this extremely easy. It will talk to your webcam and you will be able to hear the soundscapes. This will give you an idea of how you look and what else is there in the scene. In addition, to simplify matters, activate the skin filter which will only sound the exposed parts of your skin like your face assuming it is centered correctly. Try making unusual expressions and hear the soundscapes change. That way, you can determine how you are positioned.
This is independent of any AI model and there is no need to send your image to the cloud. Everything is local and your look is yours.
by Pranav ·
I had to receive my cousin at the airport. Her flight was landing at 23:45 IST. I have always been concerned about the reaction of security personnel to my video glasses. This time, I was not going inside the terminal building therefore, decided to wear my Vuzix M400 device and take a chance.
I need not have worried. No one gave me a second look or if they did, that look did not translate into any conversation or interfearance
The airport was well lit and I was able to sense people moving ahead of me. Sometimes, I saw patches of relative darkness followed by patches of light and then some kind of object like a railing or a sign. This meant that people had clustered around said obstacle and then moved away.
The OCR feature came in handy in reading the various signs.
One thing to remember at least at terminal 3 of the Indira Gandhi International Airport, is that the gate numbers are written above the gates therefore you need to raise your head to look at them unless you are tall. It is also possible to tilt the lens of the Vuzix device upwards but that did not help as much as I expected it to.
There were also times that partial signs were read. In such a case, pan your head until you get something recognizable in focus. You may also need to venture closer to the sign. It is better to frame the sign in the center of your view and then move closer.
Do watch out for people and other moving obstacles like trolleys while moving.
After reading the above pages, you may well ask, what is the point of the vOICe specially when your phone can do as good a job of text reading?
For one thing, you will need your phone to coordinate with the person and to make phone calls. who you are picking.
In addition, the visual context changes almost constantly in such an environment and in case you are moved by the crowd, you will need a mechanism to reestablish visual context. This is where an immersive setup comes in handy.
In this case, I was able to see the sign of a money exchange and was able to tell my cousin where I was standing. I did have sighted assistance with me and there were multiple money exchanges so we eventually landed up also referencing an ATM which did the trick. The sign for the ATM was harder to read probably because it was in a smaller font or may be because I was a little further from it.
So, what about those Ray-ban smart glasses? They could have helped here but as of this writing, I would need to keep queering them constantly to account for the changing context. Their things around feature would have been useful but I do not know if they would have read text that was on the objects that were all around me without sending a lot of questions to the bot. I would also not have identified the changes in landscapes as we moved between the parking and the airport terminal.
I do plan to use the vOICe inside the airport next but will see when I get the opportunity to do that.
by Pranav ·
My father died on 17 August in hospital. I am not going to elaborate on my emotions in this post.
I wanted to experience every minute of his last physical journey. He was placed on a stretcher which was probably made of bamboo. I rode with his body in the hearse. ‘
I do have my organic memories of that ride but wanted something tangible.
Thanks to Dr. Meijer and the vOICe, I was able to see his body and the flowers that it had been draped with as we drove to the crematoria.
I now have a physical and audio / visual memory of him.
Moreover, if I want to see a photograph, I have the ability to do so.
Once again, I am grateful to Dr., Meijer’s design decision that allows the vOICe to run on virtually any modern device. I did not have my video glasses with me, all I had was my iPhone 13 mini. However, in a matter of a minute, by entering the URL to the vOICe web app, I had sight.
My father was excited about AI and what it meant for me. I am placing a soundscape of the body and am placing an unedited description generated by ChatGPT above the audio element.
The image shows a close-up of a person who appears to be in a medical setting. The individual has a tanned complexion with visible signs of aging and is resting with their eyes closed. They are wrapped in white medical fabric, similar to gauze or a thin blanket, which envelops most of their head and neck area. There is a blue pad or cloth near the top of their head, possibly a part of a hospital bed or medical equipment.
The person has a white, cotton-like material inserted into their nostrils, commonly used to stop bleeding or to keep the nostrils clear. There are a few small red marks, possibly bruise or minor wound, on the visible side of their face. Their expression is relaxed, suggesting they are unconscious or sleeping.
The image’s surroundings are out of focus and not clearly defined, emphasizing the person and their immediate condition. The color palette consists mainly of blues, whites, and skin tones.
There is no visible text in the image providing additional details or context. The user interface from an image viewer is visible, showing options for editing and information, including the file name “IMG_1921.HEIC” and the file size of “1.9 MB”. The zoom level in the viewer
The image in the soundscape is without any of the screen elements mentioned in the above description. I had opened the image in the windows 11 photo viewer and had used the ai content describer plugin to get a description of the image from ChatGPT.
by Pranav ·
The Chandrayaan- 3’s landing on the moon is old news. Many of us rushed to the television to watch the live feed from ISRO.
Like most Indians, it was a proud moment for me given that landing on the south pole of the moon had not been accomplished before.
I wanted to watch a clean video and did try watching the one on youtube.
Subsequently, ISRO which I was able to watch thanks to The vOICe.
I was able to sense the change in the image as the lander got closer and closer to the lunar surface.
For best results, do the following.
Enjoy the extremely short video. Note the change in perspective as the lander gets closer.
For comparison, watch and or listen to the audio described video at the below link.
Described video of the Chandrayaan-3 landing
Note:
The video is in Hindi.