Progress On BOINC Parallel Implementation of my Lab's Brain Model

in #gridcoin7 years ago (edited)

My apologies for having been inactive for most of the last month - this was mainly a result of real life commitments catching up with me. Most significantly, I am rapidly moving towards the end of my PhD and there is still so much work I want to do - most of it a pipe dream I will never have the time for. That being said, I have made significant progress in discretely packaging simulations of my lab's brain model, with the ultimate goal of setting up a BOINC server for our lab.

For those that don't know me yet, my PhD is focussed on modelling the cells in the brain to such a level of detail that we can simulate brain function digitally. In my specific case, this allows me to observe the effects of neurodegenerative diseases such as Alzheimer's Disease to enable and advance drug development. The primary problem I encounter here is insufficient access to compute power to simulate my models to a satisfactory level of detail. BOINC will hopefully solve that limitation.

Unfortunately, no solution is perfect, and BOINC is not well suited to simulations that step through more than one dimension - in my case through space and time. In space, I model the smallest functional unit of the brain, and then connect them all together into what is called a space filling H tree, like below. Each dot is a single instance of the model. If that was the entire compute problem, I could easily cut the domain up into small pieces and wrap them in Docker wrappers to generate BOINC work units.

The problem arises when trying to step through time. Before the model can move forward in time, each node in space needs to exchange data with its nearest neighbours regarding ionic fluxes, blood volume, etc. In other words, all work at time t=1 needs to be completed before the simulation can progress to time t=2. Given that one simulation takes on the order of 100 million time steps, sending out 100 million tiny work batches, each of which needs to be returned in full before the next can be sent, is not a feasible solution.

To remedy the situation, I have attempted to vastly increase the size of the timesteps used in the simulation. This results in significant loss of accuracy, but that can be counteracted by careful interpolation to predict the behaviour between timesteps that were not explicitly simulated. What this means is that instead of stepping from t=1 to t=2 to t=3, we can step from t=1 to t=10 to t=20. Until recently the performance of large timestepping was not satisfactory, but we have just achieved successful replication of experimental BOLD data using the new implementation.

Allow me to begin by introducing the experimental data, which requires a little background knowledge. Your blood contains hemoglobin, which is used to carry oxygen to the cells that need it. Oxyhemoglobin (HbO2) is carrying oxygen, while deoxyhemoglobin (HbR) has deposited its oxygen where it is needed and is returning to the lungs to get more. We often refer to the total hemoglobin in the blood as HbT.

What you are seeing below is the cerebral vasculature of a rat after the region of the brain being observed is stimulated. I will not go into detail, as it's not pleasant how these experiments are done, and they are not done by my lab. The key thing to notice is that as a result of stimulation the blood vessels of the rat dilate (get wider), resulting in more oxyhemoglobin arriving at the active cells, which also increases the total hemoglobin present. Meanwhile, due to the significantly increased bloodflow, the deoxyhemoglobin is flushed out rapidly to return to the lungs. The time series in the bottom right shows the average behaviour observed in each of the other 3 fields over time. Note the rise in deoxyhemoglobin and drop in oxyhemoglobin at the very start of stimulation before the system corrects, which has not been simulated accurately before.

experimentalHb.png

We attempted to replicate this data using our homegenous model with large timesteps, resulting in the image below. Due to our model not capturing individual blood vessels, but the aggregate response, the first three images will look different. The time series of the aggregate behaviour is the key point of note:

modelHb.png

Not only were we able to accurately replicate the governing behaviour of the response, but we also replicated the transient dip observed at the very start of stimulation. I can imagine this does not excite any of you nearly as much as it does me, but this brings us significantly closer to encapsulating our model in a BOINC project, which in turn brings us closer to addressing neuro-degenerative diseases afflicting our elderly.


While you are not yet able to contribute to my project, there are already many other projects that can use your help! You can advance their work by donating the idle cycles of your computer, phone, tablet, you name it to the cause. Applications cover the fields of mathematics, medicine and astrophysics, allowing you to help do anything from mapping the Milky Way to developing drugs for cancer patients. To get started, you can get set up with BOINC here.

Just like Steemit rewards content creators for their work in the cryptocurrency STEEM, you can be rewarded for the research you contribute through BOINC in GRC. If you would like to be reimbursed for your contribution to academia, you can get started here.


Avatar.gif

Sort:  

I wonder if you've heard of the late James Albus, who wrote about hierarchy in the brain back in the late 70s. He did a 3 part series on his CMAC theory in Byte magazine, and there is a scan of it here: https://archive.org/stream/byte-magazine-1979-08-rescan/1979_08_BYTE_04-08_LISP#page/n67/mode/2up Your first picture reminded me of his work. He was modelling the cerebellum if I recall correctly.

That was a really good read - and no, I was not familiar with that work before. It appears that James Albus was aiming to generate a model of the brain's high level functioning such as decision making and pattern recognition, whereas we aim to model the brain's ionic functioning on a sub-cellular level.

Nevertheless, I will be digging into his models some more. Thank you for the link!

Woah! This was one hell of a post! Not often I find myself wanting a long steemit post from beginning to end and then follow up with some googling on the subject :p

Well done!

I hope I didn't force too much Googling! =)

Is it basically a cellular automaton?

  • how many nodes (cells?) do you use in your model?
  • what hardware have you used for above simulations?
  • how long it takes to process one step (generation) on your hardware?
  • computation is CPU or GPU based?

So many questions ;)

Well, it was worth waiting for such a high-quality comeback post! Glad to see it rewarded adequately. Needless to say, this is great news, being able to encapsulate such complex models into distributed computing, it will be a significant step ahead for BOINC.

Thank you for the endless feedback and support Vortac. =)

I'm looking forward to seeing the kinds of problems BOINC volunteers will be solving for humanity a decade from now. Exciting times!

Not gonna lie, I didn't under a goddamn thing. But I'm thinking I should start BOINCing things?

As far as I'm concerned, everyone should! The world's scientific community is being retrained and held back by a chronic lack of computational power, and BOINC is a platform for the public to solve that problem. If you have a computer, of any kind, you can help.

If you would like to learn more about the projects that are currently looking for volunteers, here is an exhaustive list.

I think I have a spare laptop with a decent gpu that I could start BOINCing with.

A decent GPU opens up a lot of additional projects that just CPUs won't do much for. You could, for example:

  • Decode remaining Nazi WWII Enigma messages
  • Map the Milky Way
  • Search for primes
  • Model near Earth asteroids

and a whole bunch more!

Sounds pretty sweet, I'll try it out after work tomorrow. Thanks for the info!

Hi @dutch, great post and good news for BOINC in general ! lets see what we can do together when we are pulling at the same rope into the same direction !
Iam in :)
Kind regards
Jan

The problem arises when trying to step through time. Before the model can move forward in time, each node in space needs to exchange data with its nearest neighbours regarding ionic fluxes, blood volume, etc. In other words, all work at time t=1 needs to be completed before the simulation can progress to time t=2. Given that one simulation takes on the order of 100 million time steps, sending out 100 million tiny work batches, each of which needs to be returned in full before the next can be sent, is not a feasible solution.
It is a well known problem for distributed computing. You could send N times each workunit to workers so each slice should be returned at least once pretty fast. Although it could slow down overall progress but then you don't depend on slowest worker in your swarm.

Fair call, but I am wary of redundancy. Some redundancy is accepted by a volunteer community, as we see with many BOINC projects already which verify work unit results by running each work unit twice. I doubt many people would want another layer of redundancy added on top though...

Way above my pay grade but read and upvoted nonetheless!

keep up the good work . best of luck for your Phd

Thank you very much!

loving your work, keep it up!

Oh, I didn't know your calculations were entirely (?) made through BOINC. How much CPU power do you actually need? Have you tried something like the LHC grid (that is available for some external projects in some labs)?

Currently, they are not. The intention is to port the entire model to a point where it can be distributed through BOINC.

I have considered an umbrella project, but see no need for it. Once I have work units functioning as intended, setting up a BOINC server takes all of a single afternoon.

How much CPU power do you actually need?

This is hard to answer. More compute affords me a higher resolution in the same timeframe.

I see. Thanks! Just to repeat: this is a damned good project and example of science on BOINC :)

Nice post, happy finish for your PhD !
Can you send us the link to it when ready, I would like to read the intro to try to understand a bit of it ;)
I feel both the bandwidth and latency issues (in the feedback loop). GPU can probably help, but it looks like a hell on finding a good partitioning and some hierarchy inside this stuff.

I'll post a link to my thesis when I publish, for sure!

The main problem with discretising is that if one volunteer host goes offline, I would have to wait for them to come back online and finish their assigned work before the rest of the swarm can continue. It's not ideal, but the workaround it getting there.

Ah ok, it will be nice to have a kind of replication and failover scheme (may be not easy to do, or too big data)

This sure is interesting....

very interesting your description, hope you quickly pass the program in Phd you, hopefully you quickly pass the program

Thank you - only 10 months to go!

you are wellcome

Ten months and then what? You stay at the Uni to continue working on this simulations or you are going to leave to some company?

This is extremely interesting, would love to know how you get on and if you could overcome some of the problelms. Thank you for sharing an excellent post with us.

You're very welcome, and thank you for the excellent feedback. I'm sure I will overcome the remaining hurdles in due time.

Really nice work,me being a medico is able to understand a bit of it but not completely 😁 but still i wish you continue your nice work in the research and some day that might be helpful to the humanity.

Thank you very much! I am sure we will have real life applicability of our work in the very near future.

good post! thank you for sharing a good story, success always! take a moment to see my post, if possible !!!
https://steemit.com/steem/@alfa-good/how-is-steem-power-and-steem-dollars-in-steemit-steem-is-a-liquid-currency
Thanks!
@dutch
@blocktrades
@originalworks
@good-karma
@jesta

The @OriginalWorks bot has determined this post by @dutch to be original material and upvoted it!

ezgif.com-resize.gif

To call @OriginalWorks, simply reply to any post with @originalworks or !originalworks in your message!

To enter this post into the daily RESTEEM contest, upvote this comment! The user with the most upvotes on their @OriginalWorks comment will win!

For more information, Click Here! || Click here to participate in the @OriginalWorks writing contest!
Special thanks to @reggaemuffin for being a supporter! Vote him as a witness to help make Steemit a better place!

Thankyou for sharing @dutch . . I like your post and follow next post. . I have followed you

Glory be to the Creator

thanks for sharing friends @dutch

I wish I was a blockchain developer. This makes an excellent case for something like a BrainCoin that could be a custom coin based on your project. Something precisely tailored to your design.

Good to see you guys making progress.

Really good read! I dealt with various simulations in the past and time is always the most important thing to concider. Interpolation can reduce accuracy and sometimes even completely change the outcome, but looks like you are on track in order to keep these side effects as small as possible. :)
Time steps AND distributed computing are a whole new level of complexity I wasn't really aware of until now. Really nice post, thanks you! :)

PS: I am looking forward to crunch some workunits for you in the future!

For those that don't know me yet, my PhD is focussed on modelling the cells in the brain to such a level of detail that we can simulate brain function digitally. In my specific case, this allows me to observe the effects of neurodegenerative diseases such as Alzheimer's Disease to enable and advance drug development. The primary problem I encounter here is insufficient access to compute power to simulate my models to a satisfactory level of detail.

Wow, what? That's incredible. Allthough I don't believe in the 'Big Pharma single molecule patent approach' this is really awesome I think. Being impressed by all the machine- and deep learning technologies I applaud that. Modelling a whole brain like that I never heard of before and I guess is really interesting. Would love to see that in action.

I've read an article about autism and how it affects or rather the lack of pruning of the neural network. So if you think about Rain man the classic movie, if I can take him as an example of a type of autistic case, this lack of pruning means they record everything and all the paths stay there. There's actually no highway, if you will being made, sort-of "this is the way to think or solve this problem", which I think work in real life.

Well I hope I'm not creating waves with these kind of viewpoints.