Friday, October 26, 2012

government as a design problem

trying to work on a paper revision due sooner and sooner, but i keep thinking about politics. of course, it is The Time to think about politics, but i wish i could escape it.

anyways, here's what i've been thinking, a tiny idea:

from the institution of a new state or government, for a time, it is reasonable to expect the government to grow and acquire new features. this is just because upon its institution, the government must be incomplete or flawed. virtually nothing complex can approach perfection, especially in its first design.

however, at some point, we might consider the institution - or, and here's the real idea, a given version of the institution - to be complete. that is, we have this complex structure, with many parts and many layers and many functions, and it is intended to accomplish many things under particular constraints. presumably, changes made to this structure over time are intended to fulfill these intentions. we could think of this as efficiency, i.e., how much of what the system is meant to do is it actually doing? this is a funny idea, since it implies that if the system exceeds its mandate, it is being overly efficient. i will get back to this in a moment.

the idea is that a version of the institution can be considered complete, in that a time will come when it's clear to everyone that no changes, or only basic maintenance changes, are necessary to meet the objectives of the system; or, it might be decided that the objectives are outdated, and that new objectives have arisen, and that a new system needs to be designed to replace the old one. have we ever reached that point with the american federal government? i think maybe we have, and it was a long time ago: pre-civil war, really. in the 1850s, the federal government wasn't really creating many new responsibilities for itself, and was instead preoccupied with its intended functions of maintaining relations between the states, applying tariffs in international trade, occupying new territories that would eventually become states, etc. i think this is the tail end of what historians refer to as the "second party system": FED2.0. FED2.0 was rolled out in the 1830s, had some successes early on, and then crashed and burned.

it was around the time of the civil war that the government basically went through a big redesign, acquiring new responsibilities which then required new features to be fulfilled. this was the "third party system" that lasted until the 1890s, when it was replaced with FED4.0, which lasted until the great depression. versions 3 and 4, i think, are not really considered to be very good versions (and probably could be collapsed into subversions of FED3), while a lot of people are clearly very nostalgic for versions 1.0 and maybe 2.0 (and might see those as subversions of FED1).

in the 1930s, the government went through a huge redesign: FED5.0; the end of the 1960s saw a big advance on this (FED5.1), and now we're probably at version 5.3 or 5.4. version 5 is the longest-lived political system that the US has had (or similar with FED3/4). clearly, i think, it's time for a redesign. at this point, the two parties are just concerned with adding, subtracting, or modifying features, with a strong tendency towards addition (the 'ratchet effect' or 'featuritis'). i think that a lot of people thought that with o* and the d*s, after the 2008 election, we would be moving on to a new version 6; a lot of people thought that in 2004 with b* and the r*s. neither succeeded; i don't think that either really succeeded in moving a new subversion, either: we're stuck in beta, at 5.3.2 or something like that.

so, back to 'excess efficiency'. what is that? it's not what it sounds like. when a system isn't quite fulfilling its promised aims, if it wants to preserve itself (consider that institutions don't want to die), it might throw up new proxy aims. it can them give the illusion of accomplishment or fulfillment by moving to meet those new aims, thus obscuring the fact that the old aims aren't exactly complete; or, that they're no longer valid, and that the system thus is working to fulfill aims that no longer exist. i.e., excess efficiency is a sign that a system is desperate and needs to be replaced.

Monday, October 22, 2012

taxes and politics

a friend posted this link on facebook, along with a quote to the effect of, "raising taxes to pay for investments in the middle class creates jobs". i resisted posting a response there because i don't like to argue about these things and would rather keep my opinions to myself, and because (relatedly) i am afraid to affect others' opinions of me in ways that i don't have close control over. so i thought i'd post a response here, where no one can read it:

***

while i agree generally with sentiment that says the rich should be taxed relatively more, the idea that this is then turned around by the federal government into "investment in the middle class" is not obvious to me. i think that very little of taxation, at least in a developed country like the US, translates directly into economic growth. in fact i think it tends to be the opposite, and i agree more with the idea that raising taxes tends to suppress growth.

a large portion of government spending put in place by the democratic party (which is presumably the favorable political dimension for the approving audience of this talk) is in the form of political favors to constituencies that have an insignificant impact on the economy (poor, elderly); institution of new bureaucracies which have to be funded at the same time that their mandate is usually to impose some form of *restriction* on certain types of economic activity; increased funding of dysfunctional programs without improving function ("education"); and, you can be sure that the more the US government collects in taxes, the more it will spend on the military, or on foreign aid, etc.

interesting case in point relevant to our livelihoods: significant government spending goes into biomedical research, which winds up making people live longer at the same time that it makes all forms of healthcare more expensive (MRI for everybody! one-of-a-kind cancer drug for my grampa! YOU CAN LIVE FOREVER NOW). i guess this creates jobs in the hospital/rest home industries.

*not that any of these things are wrong per se*, but it's not clear how any of this works as investment in the consumer class that generates net jobs. i think the government of a developed country actually has very little capacity to "create jobs", except in managing trade policy and maintaining transportation infrastructure. i.e. i think this guy's argument is pretty arguable.

Sunday, October 14, 2012

lull

finally working on that blur adaptation paper. not much else to think about or report.

***

last saturday (the 6th) woke up with a headache, fairly painful: i'd rate the usual ones at 2/10, and this was a 4/10. today, i wake up with one even worse: i'd put this at 6/10 (assuming it can get much worse). this is awful - it's as bad as that night tukrong punched me in the head 10 times. what is going on with my brain?

Friday, October 05, 2012

task done

we made it: the proposal is submitted. transitioning to revising papers. woo hoo?

Wednesday, October 03, 2012

inevitable negativity


i am now (technically) a faculty member at harvard medical school ("instructor": about as junior junior junior as possible)! wow! and, (technically), no longer a postdoctoral fellow: i am now a "senior scientific associate". all so i can apply for a grant with a less than a 1% chance of getting funded (optimistic i am). so, something there. beautiful appearance of progress.

also, couple of papers accepted; probably will have a paper in PLoS-one, which is nice, but i'm third of four authors, so..

***

came up with this on the train, coming home last friday night (9-28-12)

on Cambridge Street
put away
your umbrellas
or they'll wind up
cast aside
in tattered heaps
abandoned

on Cambridge Street
thrust your head
into the wind
and bear the rain
it's autumn
it's not cold yet
be thankful

Sunday, September 30, 2012

中秋快乐!

quick notes for the end of september:

week 1 of bring-your-laptop-to-work was a success; worked steadily in the lab every day, and came home each night to do particular jobs by hand, with pen and paper. extremely effective. laptop came back home friday night; going to continue this for the foreseeable future. should make the next MS revision and the following MS submission much easier.

headache last night, gradual onset; eventually focused pain above right eye socket; photophobia; went to bed, closed eyes, weird eigenlicht flicker, maybe 40-50Hz; what is that? slight headache remnant now, indistinct.

recent weirdness with reading text, usually notice in the morning; right now, left of fixation feels scotoma-like, but i can see there..

**

also, a story: when i sit at the kitchen table, in the chair by the window, i have a view of the pantry area, with the fridge and the back door. my leather sandals are wedged between the fridge and the wall, by the door, so i can wear them outside when i go to throw trash out.

i regularly mistake the sandals, peripherally, for Olive the Cat, sitting by the back door, wanting to go out. then i foveate them, and see that they are my sandals. this has happened repeatedly, maybe dozens of times: deja trompé!

Tuesday, September 25, 2012

why do i keep writing poems

Batten down the hatches!
In this electric squall
Or else we'll be sent to the deep -
The web will drown us all.

So home I'll go! To printed word,
With pen and paper work.
No opportunity to drift
Through forums or to lurk

In hiding from my calling,
I'll forge ideas by thought
And stare down syntax, words reform
To make all logic-wrought.

So batten down the hatches!
And keep the ship afloat
For though I'll try to steer us,
The net may wreck this boat.

Friday, September 21, 2012

grant, presentation, paper, model

Been trying to skip between several jobs: grant proposal with a looming deadline, modeling experiments for a paper revision with a looming deadline, looming conference presentation... well, the conference is over, and the grant is coming along, though I still do not believe I will make it.

The paper..  okay, another paper: poked an editor yesterday, and he came back with a 'minor revision' request, which I fulfilled by late afternoon today. So, finally, we have a journal article - in a 1.0 impact factor journal - to show for a 3 year postdoc. Sigh. Another in revision, in a better journal, but that's the big problem: I'm doing all these model tests, but I can't get any real momentum because I keep flipping back to the grant. Sigh. I keep complaining about the same thing. Need to set a deadline - 3 more years? - after which if I'm still making the same complaint, something needs to change.

Let's talk about the model stuff. I've talked about it already in the past few posts: in the original paper, I proposed a modification to an existing model, a minor modification, which was able to closely fit our data, but which was a bit complexified, and difficult to explain exactly why it worked as well as it did, and also unable to show how varying its parameters explained the variance in our data, etc. So, it "worked", but that's about all it did. It didn't explain much.

The existing model we call the "simple model". The simple model is indeed simple. It's so simple that it's almost meaningless, which is what frustrates me. Of course it's not that simple; you can interpret its components in very simplified, but real, visual system terms. And, it basically can describe our data, even when I complexify it just a bit to handle the extra complexity of our stimuli. And this complexification is fine, because it works best if I remove an odd hand-waving component that the original author had found it necessary to include to explain his data. Only... it doesn't quite work. The matching functions that make up the main set of data have slopes that are different in a pattern that is replicated by the simple model, but overall the model slopes are too shallow. I spent last week trying to find a dimension of the model that I could vary in order to shift the slopes up and down without destroying other aspects of its performance..  no dice.. fail fail fail.

So, I'm thinking that I can present a 'near miss': the model gets a lot of things right, and it fails to get everything right for reasons that I haven't thought hard enough about just yet. I really need to sit some afternoon and really think it out. Why, for the normal adaptor, is the matching function slope steeper than the identity line, but never steep enough? What is missing? Is it really the curvature of the CSF? How do I prove it?

Now, out of some horrible masochistic urge, I'm running the big image-based version of the "simple model". This version doesn't collapse the input and adaptation terms into single vectors until the 'blur decoding' stage. It seems like, really, some version of this has to work, but it hasn't come close yet. Looking at it now, though, I see that I did some strange things that are kind of hard to explain... Gonna give it another chance overnight.

Sunday, September 16, 2012

too far, too far

one hundred words
in haiku form
while waiting for
my flight on a
sunday evening
in september:

Rochester airport
September Sunday evening
me and three women

80s pop radio
electric piano solo
fluorescent lighting

now another man
sneakers, backwards baseball cap
the sun is setting

PA announcement
the guy's voice croaks like Stallone
a fine disco beat

two smartphones, a book
two pair boots, one pair flip flops
not a conjunction

what will our plane be?
CRJ, Boeing, Airbus?
another man comes

three women, three men
the humans are trickling in
going to Boston

the sun sets slowly
slower than it usually does
suspicious liquids

dinner of junk food
reflection of ceiling lights
in my laptop screen

Saturday, September 15, 2012

morning aura

in rochester for the OSA vision meeting.

woke up this morning about 6:30ish, with terry yelling at me to wake up. went to take a shower, and while there, realized i couldn't see my fingertips as i was washing, grabbing soap, etc. got out of the shower, got dressed, left the room and went to the lobby. got there a little after seven, and the scotoma was well into the periphery, flickering etc. it was just like the last three: left field, straight right-left through the upper field and arcing downward into the lower field.

the last one i managed to record, back in june, had a time gap between the foveal scotoma and the peripheral arcs, which i had post hoc explained as me needing to recalibrate some part of the perimeter or something. but during the last two, i noticed that the scotoma actually does seem to disappear between the foveal appearance and the peripheral arcs. wonder what is going on with that...

didn't notice any peripheral rough spot this time, but it was so early that i might just have been too dazed.. these morning scotomas that i've experienced - the last one a few weeks ago, and the one last year in the winter - they seem to have started just as i awoke. may be coincidental, since it's just a sample size of three, but i haven't had one start a half hour after waking, and i haven't woken up halfway through one (though the first time, i think i lay there with my eyes closed for the first 10 minutes or so). might be interesting to look up what sort of neurochemical changes occur in cortex, esp. visual cortex, during waking.

headache was ok, took some tylenol this time. nauseated all day long.

**

yesterday (or maybe thursday night, not sure), i remember feeling suspicious that something might be about to happen: i had the thought, i should keep track of these suspicions, to see if they're actually correlated. it is possible that i am suspicious very frequently, and just notice the coincidences..

Thursday, September 13, 2012

Deja Trompé

When I was in graduate school, I lived in Old Louisville, and walked, most days, down 3rd street to campus. Whenever I crossed the big road separating the neighborhood from campus, Cardinal Avenue, at a certain spot, I would see something up in my right peripheral visual field, and think, "Starlings!"

It was never starlings. It was always the tattered insulation hanging off a bunch of power lines strung over Cardinal. I remember this because even though I learned, pretty quickly, that I wasn't seeing starlings in that instant whenever it occurred, the fastest part of me - whatever part just automatically identifies salient stuff in the vast periphery - always thought that I was.

"Starlings!"

It not that I was hallucinating starlings. A bunch of speckly black stuff fluttering against the sky kind of looks like birds, even when you know it isn't. You can't blame me. I don't blame my visual system. It's an honest mistake. The interesting thing is that I kept making it, over and over again, with apparently no control over it. An inconsequential and incessant perceptual mistake.

I've noticed similar situations over the years, but right now I can't remember the others. I should start making a list. I bring this up because recently someone cleaned out the shared kitchen on this side of the institute, and because I always turn the lights out in the same kitchen.

I think that, because I always turn out the light when I leave the little kitchen, other people have started following my example, and now, often, when I go to the kitchen to get hot water for my tea, the light is already out. This makes me happy. It's happened very gradually. Change is slow, usually.

Usually. Recently, the development office got a new temp who is apparently a complete OCD clean freak. It's great. She cleaned this kitchen and the other one. She put up little signs everywhere telling people not to be such pigs. I love her.

Anyways, now, when I go into the little kitchen to get my water, I stand at the dispenser, watching it to make sure my hand doesn't stray and I don't get scalded, and the microwave with its little sign sits down in my lower left field. Often, lately, the light is out when I get there. I leave it that way, because there's enough light trickling in from the hallway. Every time I am in this configuration, with the light out, it looks for all the world that there is light coming out of the microwave window.

This happens over and over again. It's very robust; I can stand there and look straight at the microwave and its little paper sign, and that's what I see; then I look away, and the sign becomes an emission of lamp light from within the microwave. I can turn the mistake on and off by moving my eyes back and forth.

Again, I don't blame my visual system. It's doing the best it can. I've seen so many microwaves, and when they're cooking, they usually have little lamps inside, so you can see your whatever rotating on the little turntable. If the room is dark, the image is basically of a luminous rectangle in the front door of a microwave. Not many microwaves that I have known have worn little paper signs on their doors. To their disgusting, disgusting peril.

There must be a name for this, but I can't find it. So for now I'm going to invent a term: deja trompé‎, "fooled again". Deja as in deja vu, "again seen"; trompé‎ as in trompé l'oeil, "deceives the eye". Seems like the right flavor for this sort of thing. I'll start keeping track of these, however rare they are. I'll inaugurate the list with a new entry label.

BACK TO WORK

Wednesday, September 12, 2012

adaptomatic

Trying to figure out how to proceed with this adaptation paper, and I retreat here.

Minor problem is the rewrite: this will get done, not too worried about it. May be the last thing that gets done, since the major problem needs to be solved materially first.

Major problem is the modeling. The original paper details a complexified version of the model proposed by the authors of a paper that our paper basically replicates, accidentally. We were scooped, and so I thought that to novelify our paper, I would take their model and try to push it a little further, and do some extra analysis of it.

What I didn't do was what I should have done, which was to also test the simple model and show that it is somehow inadequate, and that complexification is therefore justified or necessary. I am actually ambivalent about this. My main idea was that we should take a model which has generalizable features and use it to explain the data; but, it's true that the more sophisticated version can't really be credited with achieving anything unless the simple one can also be shown to fail.

So the problem is that I have to do a lot of testing of the simple model. So, I decided that I would scrap the section that was already in the paper and replace it with an evaluation of the simple model, but make up for the lack of 'advance' by employing the simple model in a more realistic simulation of the actual experiments. This is what I've been trying to do, and basically failing at, for several weeks now.

The first idea was to use the simplest form of the model, but the most complete form of the stimuli: videos, played frame by frame and decomposed into the relevant stimulus bands, adaptation developing according to a simple differential equation with the same dimensions as the stimulus. This didn't work. Or, it almost worked. The problem is that adaptation just won't build up in the high frequency channels, unless it's way overpowered, which is against any bit of evidence I can think about. If high frequency adaptation were so strong, everything would be blurry all the time. I think it should be the weakest, or the slipperiest.

Soon after that, I gave up and retreated to the 'global sum' model, where instead of using 2d inputs, I use 0d inputs - i.e. the stimulus is treated as a scalar. I get the scalars from the real stimuli, and the same dynamic simulation is run. It's tons faster, of course, which makes it easier to play around with. I figured I would have found a solution by now.

See, it's so close. It's easy to get a solution, by adjusting the time constants, how they vary with frequency, and the masking strength, and get a set of simulated matching functions that look a lot like the human data. But I figure this is uninteresting. I have a set of data for 10 subjects, and they seem to vary in particular ways - but I can't get the simulated data to vary in the same way. If I can't do that, what is the point of the variability data?

Also, last night I spent some time looking closely at the statistics of the original test videos. There's something suspicious about them. Not wrong - I don't doubt that the slope change that was imposed was imposed correctly. But the way contrast changes with frequency and slope is not linear - it flattens out, at different frequencies, at the extreme slope changes. In the middle range, around zero, all contrasts change. Suspiciously like the gain peak, which I'm wondering isn't somehow an artifact of this sort of image manipulation.

I don't expect to figure that last bit out before the revision is done. But, I'm thinking it might be a good idea to play down the gain peak business, since I might wind up figuring out that e.g. adaptation is much more linear than it appears, and that the apparent flattening out is really an artifact of the procedure. I don't think I'll find that, but - did I mention I'm going to write a model-only paper after this one? - seems a good idea not to go too far out on a limb when there are doubts.

I have a nagging feeling that I gave up too soon on the image-based model...

Friday, September 07, 2012

talk: 97%


did a dry run today for my FVM talk. i think it went well, but there was a good amount of feedback. (incidentally, earlier this week i came to the lab, and passed my preceptor e* talking with a familiar old guy in the hall; a few minutes later, e* brings the guy to my office and asks me to show him my work. the old guy was l.s., one of the elder statesmen of european psychophysics. turns out he had been a postdoc at the instutute more than 40 years ago, and was in town, and had just dropped in to see old friends.. i took him through my presentation at quarter speed, and he was very enthusiastic. made some suggestions about controlling for the 'knowledge' aspect of my stimuli and experiment design. took notes. had a good talk with him, he seems to know my grad school mentor well, knows all his students. so i didn't go to ECVP this week, but i got to spend a morning with one of its founders...)

anyways, the dry run: p* was the only one, as i guess i expected, to make real comments on the substance of the talk. he had two points/questions:

1. what happens if the two images are different, i.e. if they have different phase spectra? i have not tried to do this experiment, or to predict the result. i guess that technically, the model that i am evaluating would make clear predictions in such an experiment, and the perceptual process i am claiming to occur would be equally applicable. but, really, i am tacitly assuming that the similarity of the two images is tamping down noise that would otherwise be there, somehow in the spatial summation, that isn't actually reflected in the model but that would be there for the humans. but, it might work just fine. i should really try it out, just to see what happens... (*edit: i tested it in the afternoon, and the result is exactly the same. experiment is harder, and the normalization is wacky, but seems clear it works...)

2. don't the weighting functions look just like CSFs? isn't this what would happen if perceived contrasts were just CSF-weighted image contrasts? yeah, sure, but there's no reason to believe that this is how perceived contrast is computed. the flat-GC model is close to this. i wonder if i shouldn't just show a family of flat-GC models instead of a single one, with one of them having 0-weighted GC...

the other main criticism was of the slide with all the equations. this is the main thing i think i need to address. i need to remake that slide so it more naturally presents the system of processes that the equations represent. some sort of flow or wiring diagram, showing how the equations are nested...

also need to modify the explanation of the contrast randomization; not add information, but make clearer that the two contrast weighting vectors are indeed random and (basically) independent.

Monday, September 03, 2012

two out of three ain't enough

okay, so, really, i spent the labor day weekend watching youtube videos, looking at funny gifs, reading the news, and other random things, while running half-baked model simulations for the blur adaptation revision.

first thing i did was to run the video-based model through the experiment on the same three adaptation levels used in the original experiment. it worked at an operational level, i.e. it matched sharper things with sharper things and blurrier things with blurrier things, and the effects of the adaptors were correctly ordered - it didn't do anything crazy. on an empirical level, though, it was wrong.

for the original subjects, and most of the replication subjects, the perceived normal after blank adaptation should be matched to a slightly sharpened normal-video-adapted test; the simulation did the opposite. not a huge problem, but like i said, against the trend.

bigger problem is that the simulation failed to get the 'gain' peak for the normal adaptation condition; instead, gain just increased with sharpness of the adaptor. now i'm rerunning the simulation with some basic changes (adding white noise to the spatial inputs, which i don't think will work - might make it worse by increasing the effective sharpness of all inputs - but might have something of a CSF effect; and windowing the edges, which i should have done from the start).

one funny thing: even though the gain for the sharp adaptor is too high (being higher than for the normal adaptor), the gains for the normal and blurred adaptors are *exactly* the same as the means for the original three subjects: enough to make me think i was doing something horribly weirdly wrong in the spreadsheet, but there it is:



weird, but too good to be true. undoubtedly, every change to the model will change all of the simulation measurements, and the sim is definitely as noisy as the humans - even the same one run again would not get the same values.

Sunday, September 02, 2012

random

I seem to have gotten into treating this thing as a migraine journal, so here: headache last night (Saturday). Strange one, came on slowly, from mid-afternoon, increased gradually until 10 or so, when it was actually pretty irritating. May be something else. It's kind of still here, vaguely. Front of the head, above-behind the eyes, but something about it is different. Dunno.

As for work, I should have done more this weekend. I have 3 current main foci: FVM presentation, blur adaptation revision, and R01 application.

The presentation is >90% done. I'm leaving it for a few days.

The blur adapt revision is 0% done. I'm trying to figure out what "simple" model to replace the section in the paper with. If I can't get it to work by the end of the week, I think I'll have to stick with the original "complicated" model, and *add* material (thus making it *more* complicated) to explain why the simple version can't be easily adapted to work. What this entails is about an hour of programming and 24 hours of running the simulations/measurements so I can see the results and decide on what isn't working and make changes and repeat the process. In the meantime, I do nothing productive. So:

R01 application is... well... I don't want to do it. It's futile, but it's my job. Will start soon. Should have started this weekend.

Wednesday, August 29, 2012

dream science?

Too many entries - let this be the last one for August.

Fantastically, incredibly, unbelievably, implausibly, Monday night I had a dream that directly relates to Monday's entry. I'll leave out irrelevant details: I dreamed that I was experiencing a migraine aura.

In the dream, I noticed the phosphene-like foveal scotoma, and at first had the "is it an afterimage? what bright light did I look at?" reaction, and then realized what it really was. It was upsetting, actually, because the last one was just 1 week previous, and I felt like once a week is a bit too frequent.

I then set about trying to record the aura with my perimetry program, except that my computer was now a large, flat panel lying on the floor, like a giant i-pad. The layout was of course different - not a blank gray screen, but a thin-line black grid, like a Go board, on a wood-brown background. Jingping was there, and kept trying to move the grid around, and I kept telling her to stop.

Once I was trying to record it, the scotoma was no longer foveal, but extended 10-20 degrees out, straight to the left and then arcing downward towards the inferior vertical meridian. This makes me think that I wasn't actually experiencing an aura in my sleep - to get from the foveal scotoma to 10-20 degrees should take 15-20 minutes, and I don't think that much time actually passed in the dream - it seemed like less than a minute. Of course, time and space are both funny in dreams, so who knows. There was no headache on Tuesday, anyways.

It was very frustrating trying to set the fixation point in the dream perimetry program. I just couldn't fixate - I would set it in one place, and then felt that it should be somewhere else. I think I finally gave up and started sticking my hand in the scotoma to probe its size.

So, whether or not I was really experiencing an aura, or just dreaming that I was experiencing one, is an interesting question. It seemed like a real one, and I noted lots of spatial details: the tiny phosphenic bead of the foveal scotoma, the fuzzy noisiness of the peripheral scotoma arc (though the periphery seemed clearer somehow in than true peripheral vision), the thin black lines of the perimetry grid, the unfixable fixation spot.. If visual experience includes V1 activity, and if the visual aura occurs in V1, and if V1 is quiet or suppressed during dreaming, how could I have seen what I did, unless spatial vision includes a good deal of higher-level inference?

It seems that I proposed an experiment on Monday afternoon, and then did the experiment in my sleep that night. I have never been so efficient!

Monday, August 27, 2012

summation or conclusion?

So, I'm realizing now that this note from a few days ago is touching on this entry from several months back (if only I could keep everything in my head at once...). In the latter, I was talking about the idea that visual experience is a stack of phenomena, extending all the way down to the optical image, even to the light field, and all the way up to cognition and emotion. In the former, I realized that my standing, computational interpretation of the classification image experiment involves an assumption that estimates of a particular psychological construct - perceived contrast - are mediated by the same processes whether the stimulus is simple or complex.

This stance doesn't conflict with the 'stack' idea, but when you think of both together it seems dubious. With simple stimuli, there isn't much else elicited by the visual pattern, so estimates of its properties can be localized to a small set of possible mechanisms, which is the point of using simple stimuli in the first place. So, there are multiple layers to the stack, but most of them are relatively empty or inactive. When the stimulus is complex, all those other stacks are now active, and filled with activity which is ostensibly more important and interesting to the observer. Is it reasonable to continue to assume that the observer can make use of the same information in that 'spatial vision' layer that he could when there was nothing to distract him elsewhere?

I realized this connection because I was thinking of the implications of one alternative (complex visual qualia are the result of highly nonlinear summation of simple visual qualia) or the other (complex qualia may be inferences drawn from 'basis' qualia, that could possibly exist - as perhaps in a dream - independently of those bases). How do you tell the difference? Take away the spatial vision level, and see what is left. How to do this? Lesions maybe, but the first thing that comes to mind is to compare what imagery looks like when you're seeing it versus when you're dreaming it.

Friday, August 24, 2012

punched in the head

title says it all. punched a few times in the head, and now i have a headache. feels migrainish. it is possible that getting punched in the head in just the right way can trigger a migraine. or maybe i have a concussion? i do have a big red bruise on my forehead, so maybe the pain is on the outside, muscular, and i just can't tell the difference since it's all just front-of-the-head.

anyways.

**edit, 0:23, 8/27/12
the bruise is still there, fading, and the flesh is a bit tender - worse yesterday - but the headache was gone when i woke up saturday morning.

Thursday, August 23, 2012

vacation report

Spent the last 5 days (Sat-Wed) down in Tennessee/Alabama, visiting family. Monday morning, Jingping woke up about 8 and went looking around and came back saying that my parents were still home, when she thought they should be out taking a walk somewhere. I was barely awake, still hadn't opened my eyes. When I finally did a few minutes later, I found that I was halfway through a scintillating scotoma, maybe around the 15-20 minute point. It looked a lot like last time, left field, relatively straight scotoma from above fixation leftward, arcing downward and below. I got out of bed and went to sit in the sunroom to watch the rest of it. The scintillation was rather weak, but still noticeable - I knew what was happening within a second or two of opening my eyes. The headache started soon after I got out of bed, and was kind of a bad one. Above-behind my eyes, focused on the right side. Nauseated and dizzy for a day, which sucked because I had to drive down to Huntsville Monday afternoon (in my parents' Prius with an expired Kentucky driver's license, don't tell my mother). Still hurt a bit Tuesday night.

I think that maybe the slight headache I described on the 16th might have been part of the prodrome for this one, otherwise I didn't notice any signs.

**

Last night on the way home I had an insight into how to explain the low-pass gain control that I'm proposing. A basic Barlow-Foldiak type anti-Hebbian learning rule should develop low-pass weights if a set of scaled filters is repeatedly exposed to low-pass input, or maybe even if it's just exposed to white noise. Gonna try this later today!

Friday, August 17, 2012

contrast or inference?

Norma Graham makes an interesting point, which I've seen quoted many times, in her book on spatial vision. She notes that it is as if the brain is transparent to what is happening in the early levels of visual processing, and that this is curious. It's curious, but it's a typical stance for someone who studies spatial vision; we assume that discrimination or identification or detection of signal strength is mediated almost entirely by the filters that are transducing the signal, not those that understand it, or respond (overtly) to it.

Whether or not this transparency holds when complex images are viewed is, I think, totally unknown. It may be that the percept elicited by a complex scene really is the sum of its parts, and that it simply provokes additional sensations of meaning, identity, extension, etc., which are tied to spatial locations within the scene. So, the visible scene that we are conscious of is indeed an object of spatial vision. This is the point of view I generally adopt, and I think it is common.

Another view is that the percept is entirely inference. Boundaries, surfaces, colors, textures, etc., are qualities in themselves, inferred from particular organization of spatial structure, and these then are organized in such a way that objects and identities and meanings can be inferred in successive stages. These inferences are what is seen consciously; perhaps inferences, and the evidence for them (the matter of spatial vision), are experienced simultaneously, but the substantive inferences, being the important elements of experience, are what dominate consciousness. So, only a small part of the phenomenal scene is actually constituted by e.g. luminance contrasts, and much more of it is constituted by higher level inferences. I think this point of view is also common, maybe especially in the current generation of visual neuroscientists.

The latter view is not exclusive of Graham's observation. If the patterns that are viewed are simple enough, they will not form objects, and will not have meaning. Or, they will be interpreted only as what they are, which doesn't require much inference, or only circular inference (which isn't a bad thing necessarily, when you really do want to conclude that a thing is itself, e.g. a gaussian blob of light, on the basis of its being a blob of light; usually, you want to infer that there is a letter on a page on the basis of a particular arrangement of blobs of light).

So, in the experiment that I'm currently analyzing to death, I am clearly taking the first view, in which case I think my conclusions are solid. If the second view is more accurate, what does the result mean? It could mean that inferences about image strength are based on higher frequencies just because they are the more susceptible to loss in a weak signal. If I'm asking subjects to judge image contrast, they could easily interpret this as judging image strength, and then their judgments would be biased towards the most delicate parts of the image, but they would still take everything into account.

This latter interpretation is still interesting, but it doesn't require "suppression". It is worth mentioning and I should at least include it in the manuscript, although the FVM talk probably will not have space... already there's barely space for the default story.