[visionlist] [cvnet] Trying to demonstrate stereoscopic vision remotely
loschky at ksu.edu
Sat Mar 21 23:56:41 -04 2020
Thanks, Jeremy! I'll use that! We're already using your book! Evidently
I need to read it more carefully! ;)
On Sat, Mar 21, 2020 at 5:43 PM Wolfe, Jeremy M.,Ph.D. <
jwolfe at bwh.harvard.edu> wrote:
> Hi Lester
> I think you are leaving out the most straight-forward approach of teaching
> students how to free fuse
> It works quite well with simple stimuli and a little practice.
> Here is the relevant figure from our S&P textbook.
> Good luck everyone!
> On Mar 20, 2020, at 10:45 PM, Lester Loschky <loschky at ksu.edu> wrote:
> External Email - Use Caution
> Hi Everybody,
> If you teach Sensation and Perception, and are currently preparing to
> teach it remotely, you may have the same question I have: how can you
> demonstrate stereovision remotely?
> As preface, the following are methods I have used in in-person classes to
> demonstrate stereo vision:
> 1. an actual stereoscope and example stereoimages to share with
> students (including the classic Julesz square tile random-dot-stereogram
> 2. example stereoscopic lenticular lens images to share with students
> 3. red/green anaglyph images with sets of cardboard & plastic
> red/green anaglyph glasses
> 4. Google Cardboard plus cell phone to share with students
> 5. random dot autosterographic images
> 6. touching two pen tips together using two eyes versus one eye
> 7. learning about crossed vs. uncrossed disparity using two fingers at
> different distances
> Unfortunately, my students don't uniformly have access to the apparatuses
> required for 1-4 above.
> Re. # 3 (red/green anaglyph images), I've thought of having students order
> a single pair of red/green anaglyph glasses online. However, it appears
> that the cardboard and plastic ones can only be purchased in bulk. (I guess
> they're too cheap to sell individually.) They also might not arrive in
> time, but students could still enjoy them once they get them.
> Re. #4 (Google Cardboard), I recall getting a free Google Cardboard from
> the NYTimes several years ago. However, they are now no cheaper than $5
> (Irisu, of India), and likely wouldn't arrive in time.
> Regarding option #5 (random-dot autostereograms), I have found that since
> seeing random dot autostereographic images in depth requires perceptual
> learning, a large proportion of students don't manage to learn (within the
> short time period given in a single class period). (Of course, many
> students may have a lot of time on their hands now, so they might keep at
> it long enough to learn to perceive them. But there will definitely be a
> good proportion of students who don't try long enough to learn, and so
> don't get it.)
> #6 (touching two pen tips together) is definitely something that can be
> done remotely. However, it doesn't have the "Wow!" factor of other
> demonstrations. It is more of an "oh, really..." experience to realize how
> much worse you are with one eye than two.
> #7 (using two fingers at different distances to teach crossed vs.
> uncrossed disparity) can definitely be done remotely. It is very
> educational, but again does not have the "Wow" factor.
> There is also the finger "hot dog" illusion, which can be done remotely.
> It is interesting, but quite different from all of the others in that
> stereoscopic depth perception is not involved.
> For the related phenomenon of motion parallax, "wiggle vision" is a very
> nice demonstration:
> Of course, depth perception from motion parallax is importantly
> theoretically related to stereoscopic vision (both involve two different
> images from two different views, one seen over time (and only needing one
> eye)--motion parallax--and the other seen simultaneously (and requiring two
> eyes)--stereovision). But it is not the same as stereoscopic vision, so is
> a separate but related issue.
> For the related phenomenon of binocular disparity, there is the famous
> "hole in your hand" illusion using a cardboard paper towel roll. If
> students have a spare cardboard paper towel roll, they can do this
> remotely. But, again, it is a theoretically related but separate issue.
> Any other suggestions would be appreciated.
> Best wishes,
> Lester Loschky
> Associate Director, Cognitive and Neurobiological Approaches to Plasticity
> Department of Psychological Sciences
> 471 Bluemont Hall
> 1114 Mid-Campus Dr North
> Kansas State University
> Manhattan, KS 66506-5302
> email: loschky at ksu.edu
> research page: https://www.k-state.edu/psych/research/loschkylester.html
> lab page: http://www.k-state.edu/psych/vcl/index.html
> cvnet mailing list
> cvnet at lawton.ewind.com
> Jeremy M Wolfe, PhD
> Professor of Ophthalmology & Radiology,
> Harvard Medical School
> Visual Attention Lab
> Department of Surgery
> Brigham & Women's Hospital
> 65 Landsdowne St
> 4th Floor
> Cambridge, MA 02139
> Phone: 617-768-8818
> Fax: 617-768-8816
> Best email: jwolfe at bwh.harvard.edu <jwolfe at bwh.harvard.edu>
> Backup: jeremywolfe0131 at gmail.com
> URL: search.bwh.harvard.edu
> Editor: Cognitive Research: Principles and Implications (CRPI)
> CRPI is the new open access, peer-reviewed journal of the
> Psychonomics Society
> Do you do "use-inspired, basic research" in Cognition? That is what we
> The information in this e-mail is intended only for the person to whom it
> addressed. If you believe this e-mail was sent to you in error and the
> contains patient information, please contact the Partners Compliance
> HelpLine at
> http://www.partners.org/complianceline . If the e-mail was sent to you in
> but does not contain patient information, please contact the sender and
> dispose of the e-mail.
Associate Director, Cognitive and Neurobiological Approaches to Plasticity
Department of Psychological Sciences
471 Bluemont Hall
1114 Mid-Campus Dr North
Kansas State University
Manhattan, KS 66506-5302
email: loschky at ksu.edu
research page: https://www.k-state.edu/psych/research/loschkylester.html
lab page: http://www.k-state.edu/psych/vcl/index.html
-------------- next part --------------
An HTML attachment was scrubbed...
-------------- next part --------------
A non-text attachment was scrubbed...
Size: 128418 bytes
Desc: not available
More information about the visionlist